Answer Engine Insights

Last updated: April 19, 2026

The Answer Engine Insights node pulls structured data from your Profound account directly into an Agent. Use this step when you want to analyze visibility, sentiment, and citations for your brand across AI answers, then pass that data into downstream steps such as LLM analysis, reporting, or alerts.

You can configure exactly which metrics, dimensions, filters, and date ranges to use, making this node a flexible way to operationalize Profound data.

aei.png

See this document for additional instructions on adding this node to an Agent, and this document for a full list of available nodes.


When to use this node

Use Answer Engine Insights for tasks such as:

  • Monitoring how often your brand is mentioned and cited in AI answers

  • Comparing visibility or sentiment across regions, topics, or models

  • Building automated reports on share of voice over time

  • Triggering Agents when visibility drops or sentiment changes

  • Feeding brand performance data into LLM steps for narrative summaries or insight generation


Node configuration

Selecting the Answer Engine Insights node opens its configuration panel on the right side of the Agent Builder.

Report Type (required)

Choose which type of insight you want to retrieve:

  • Visibility – Measure how often your brand appears in AI answers and its share of voice.

  • Citation – Focus on citations and mentions (e.g., how often your brand is cited, where, and by which models).

  • Sentiment – Analyze positive, negative, or neutral sentiment toward your brand in AI answers. Sentiment data is generated from sentiment prompts, which are automatically created by Profound based on your tracked topics and follow a formulaic format (e.g., "Evaluate [your brand] on [topic]"). These prompts are always branded and run for both your company and competitors (in Prompts tab, toggle to "Sentiment"). Results appear in the Sentiment tab (not the Prompts tab).

The options available in other drop downs (Metrics, Dimensions, Filters) may vary based on the selected report type.

Note: Sentiment prompts are separate from visibility prompts and count toward your total prompt limit. They can be edited, customized, or deleted in your Profound account if you prefer to allocate that prompt budget elsewhere. Without sentiment prompts, the Sentiment tab in Answer Engine Insights will not be populated.

Brand Category (required)

Select the Category you want to analyze. This corresponds to the categories configured in your Profound account (for example, specific product verticals or brand groupings).

You can select a single category per node. To compare multiple categories, add multiple Answer Engine Insights steps.

Metrics (required)

Choose one or more metrics to return in the report. This is the actual data metric that is returned by the node. Available metrics depend on the selected report type, but may include values such as:

  • (Citations) Count – Number of answers, mentions, or citations matching your configuration.

  • (Citations) Share of Voice – Relative presence of your brand compared to others within the selected category.

Each metric will be included in the node’s output.


Dimensions (optional)

Dimensions let you “slice” your metrics into smaller segments. When you select one or more dimensions, the system returns separate metric values for every unique combination of those dimensions.

Example dimensions include:

  • Date

  • Region

  • Topic

  • Asset Name

  • Tag

Example
If your metric is Visibility Score and you choose Region and Date as dimensions, the output will include one Visibility Score value per region per day. For instance:

[
{
"metrics": [6.79033503559518],
"dimensions": ["2025-12-09","United States"]
},
{
"metrics": [6.780276789866595],
"dimensions": ["2025-12-08","United States"]
},
{
"metrics": [3.63023942039234],
"dimensions": ["2025-12-08","Argentina"]
},
{
"metrics": [4.92039201029494],
"dimensions": ["2025-12-09","Argentina"]
}
]

This allows you to answer questions such as:

  • Share of voice by date

  • Count by region

  • Sentiment by model

If you do not select any dimensions, the result is a single aggregated value for each metric across the entire configuration.

Filters (optional)

Use filters to narrow the dataset to specific entities or attributes. Click Add Filter to configure:

  • Field – The field you want to filter on, such as:

    • Region ID

    • Model ID

    • Topic ID

    • Tag ID

    • Hostname

    • Path

    • URL

    • Root Domain

    • Prompt Type

  • Operator – For example, is or other comparison operators (depending on the field).

  • Value – The value to match, such as a region UUID or hostname. You can:

    • Paste a static value, or

    • Insert a value dynamically from earlier Agent steps by typing /

Multiple filters can be added to refine the dataset further (for example, a specific model in a specific region).

When you filter on a field, you must also include that field as a dimension.

Date Range (required)

Specify the time window to query. You can select relative ranges (e.g., Last 7 Days) or configure other options depending on your Profound setup.

The node will return data only for answers that fall within this range.

Output Label (required)

Provide a descriptive label for the node’s result. This becomes the variable name used in downstream steps.

Examples:

  • aei_citation_report

  • visibility_score

  • sentiment_themes


Advanced settings

Expand Advanced settings to control additional parameters.

Limit

Set the maximum number of rows returned by the query. This is particularly useful when you:

  • Only need a sample or top N results

  • Want to keep downstream processing lightweight

For example, set 10 to return the top 10 rows based on your selected metrics and dimensions.


Output

The Answer Engine Insights node returns a structured result containing:

  • The selected metrics

  • Any dimensions used for grouping

  • Values constrained by your filters and date range

The output can be consumed by:

  • Prompt LLM nodes (for narrative summaries, recommendations, or anomaly detection)

  • Additional transformation steps

  • Reporting or alerting Agents


Example usage

1. Monitor brand visibility by region

  1. Add an Answer Engine Insights node.

  2. Set Report Type to Visibility.

  3. Choose the relevant Brand Category.

  4. Select Share of Voice under Metrics.

  5. Choose Asset and Region and Date as Dimensions.

  6. Set Date Range to Last 7 Days.

  7. Filter Asset name to your asset name

  8. Optionally, set Limit in Advanced settings to control the number of rows.

  9. Use a Prompt LLM step to summarize regional performance:

Summarize the following visibility data by region and highlight any notable changes or outliers:

{{visibility_insights}}

2. Analyze sentiment by model

  1. Add an Answer Engine Insights node.

  2. Set Report Type to Sentiment.

  3. Choose your Brand Category.

  4. Select a sentiment-related Metric (e.g., counts or ratios by sentiment).

  5. Add Model and Asset as a Dimension.

  6. Filter Asset name to your asset name

  7. Optionally, filter on Region ID to narrow the analysis.

  8. Use the output in an LLM step to generate a narrative summary for stakeholders.

3. Pull top citation domains

  1. Set Report Type to Citation.

  2. Select Count as the Metric.

  3. Use Hostname as a Dimension.

  4. Set Limit (e.g., 10) to retrieve the top citing hosts.

  5. Feed the output into downstream steps for outreach planning or competitive research.


Best practices

  • Start broad (few filters, key metrics) and refine as needed to avoid overly narrow results.

  • Use Dimensions for the breakdowns you care most about (e.g., region, model, topic) and keep the set small for easier downstream interpretation.

  • Use Limit to keep responses manageable, especially when you plan to pass data into LLM steps.

  • Use clear, descriptive Output Labels so your Agents remain understandable as they grow.


Troubleshooting

The node returns no data

  • Confirm that the selected Brand Category, Date Range, and Filters match data you expect in Profound.

  • Try removing filters or expanding the date range to validate that data exists.

Results are too granular or long

  • Reduce the number of Dimensions.

  • Lower the Limit in Advanced settings.

Variables are not being applied in filters

  • In the filter value field, type / and select from the variable picker rather than pasting raw placeholders.