Skip to main content

Mental Models

User-curated summaries that provide high-quality, pre-computed answers for common queries.

What Are Mental Models?

Mental models are saved reflect responses that you curate for your memory bank. When you create a mental model, Hindsight runs a reflect operation with your source query and stores the result. During future reflect calls, these pre-computed summaries are checked first — providing faster, more consistent answers.

Why Use Mental Models?

BenefitDescription
ConsistencySame answer every time for common questions
SpeedPre-computed responses are returned instantly
QualityManually curated summaries you've reviewed
ControlDefine exactly how key topics should be answered

Hierarchical Retrieval

During reflect, the agent checks sources in priority order:

  1. Mental Models — User-curated summaries (highest priority)
  2. Observations — Consolidated knowledge
  3. Raw Facts — Ground truth memories

Mental models are checked first because they represent your explicitly curated knowledge.


Create a Mental Model

Creating a mental model runs a reflect operation in the background and saves the result:

# Create a mental model (runs reflect in background)
result = client.create_mental_model(
bank_id=BANK_ID,
name="Team Communication Preferences",
source_query="How does the team prefer to communicate?",
tags=["team", "communication"]
)

# Returns an operation_id - check operations endpoint for completion
print(f"Operation ID: {result.operation_id}")

Parameters

ParameterTypeRequiredDescription
namestringYesHuman-readable name for the mental model
source_querystringYesThe query to run to generate content
idstringNoCustom ID for the mental model (alphanumeric lowercase with hyphens). Auto-generated if omitted.
tagslistNoTags that scope the model during reflect and filter source memories during refresh. Defaults to all_strict matching, so only memories carrying every listed tag are read. See Tags and Visibility.
max_tokensintNoMaximum tokens for the mental model content
triggerobjectNoTrigger settings (see Automatic Refresh)

Create with Custom ID

Assign a stable, human-readable ID to a mental model so you can retrieve or update it by name instead of relying on the auto-generated UUID:

# Create a mental model with a specific custom ID
result_with_id = client.create_mental_model(
bank_id=BANK_ID,
name="Communication Policy",
source_query="What are the team's communication guidelines?",
id="communication-policy"
)

print(f"Created with custom ID: {result_with_id.operation_id}")
tip

Custom IDs must be lowercase alphanumeric and may contain hyphens (e.g. team-policies, q4-status). If a mental model with that ID already exists, the request is rejected.


Automatic Refresh

Mental models can be configured to automatically refresh when observations are updated. This keeps them in sync with the latest knowledge without manual intervention.

Trigger Settings

SettingTypeDefaultDescription
mode"full" | "delta""full"Refresh strategy. See Refresh Mode below.
refresh_after_consolidationboolfalseAutomatically refresh after observations consolidation

When refresh_after_consolidation is enabled, the mental model will be re-generated every time the bank's observations are consolidated — ensuring it always reflects the latest synthesized knowledge.

Refresh Mode

Two strategies are available for how a refresh produces the new content:

  • full (default) — every refresh regenerates the entire content from scratch. Simple and predictable: the LLM synthesises a fresh document from the retrieved memories. Best when the document is short, when you want every refresh to potentially restructure the output, or when you're not yet sure what the final shape should be.

  • delta — refresh emits a list of typed operations (add a section, append a bullet, replace a block, remove a stale paragraph) against the document's existing structure, then renders the result. Sections that aren't targeted by any operation are copied through byte-identical — no paraphrasing, no whitespace drift, no list-style normalisation. Best for long-lived "playbook"–style mental models where you want stability across refreshes and only the genuinely changed parts to move.

Delta mode falls back to a full regeneration automatically in two cases:

  1. The mental model has no existing content yet (nothing to anchor edits on).
  2. The source_query has changed since the last refresh (the topic has shifted; the existing structure may no longer apply).

If the LLM call fails or returns an empty answer, the existing content is preserved — refreshes never overwrite a populated document with an empty one.

Use CaseRecommended ModeWhy
Skill / playbook docsdeltaSections live for many refreshes; only specific rules change
Onboarding summariesdeltaAdding new team members shouldn't restructure the doc
Real-time dashboardsfullEach refresh is a fresh snapshot
Short FAQ summariesfullWhole-document regeneration is cheap and unambiguous
# Create a mental model with automatic refresh enabled
result = client.create_mental_model(
bank_id=BANK_ID,
name="Project Status",
source_query="What is the current project status?",
trigger={"refresh_after_consolidation": True}
)

# This mental model will automatically refresh when observations are updated
print(f"Operation ID: {result.operation_id}")

When to Use Automatic Refresh

Use CaseAutomatic RefreshWhy
Real-time dashboards✅ EnabledStatus should always be current
Policy summaries❌ DisabledPolicies change infrequently, manual refresh preferred
User preferences✅ EnabledPreferences evolve with new interactions
FAQ answers❌ DisabledAnswers are curated, should be reviewed before updating
tip

Enable automatic refresh for mental models that need to stay current. Disable it for curated content where you want to review changes before they go live.


List Mental Models

# List all mental models in a bank
mental_models = client.list_mental_models(bank_id=BANK_ID)

for mental_model in mental_models.items:
print(f"- {mental_model.name}: {mental_model.source_query}")

Get a Mental Model

# Get a specific mental model
mental_model = client.get_mental_model(
bank_id=BANK_ID,
mental_model_id=mental_model_id
)

print(f"Name: {mental_model.name}")
print(f"Content: {mental_model.content}")
print(f"Last refreshed: {mental_model.last_refreshed_at}")

Detail Levels

Both List and Get endpoints accept an optional detail query parameter that controls how much data is returned. This is useful for reducing response size, especially in agent boot flows or MCP clients where context budget is limited.

LevelFields ReturnedUse Case
metadataid, bank_id, name, tags, last_refreshed_at, created_atInventory — "what models exist?"
contentAll metadata fields + source_query, content, max_tokens, triggerAgent boot — "what do the models say?"
full (default)All fields including reflect_responseDeep inspection — "what evidence backs this model?"
# List only names and tags (smallest response)
curl "$BASE_URL/v1/default/banks/$BANK_ID/mental-models?detail=metadata"

# List with content but without provenance chains
curl "$BASE_URL/v1/default/banks/$BANK_ID/mental-models?detail=content"

# Get full detail (default behavior)
curl "$BASE_URL/v1/default/banks/$BANK_ID/mental-models/$MODEL_ID?detail=full"

The detail parameter is also available in the MCP tools:

{"bank_id": "my-bank", "detail": "metadata"}
tip

Use detail=content for agent orientation flows. It includes everything the agent needs to understand the models without the heavyweight reflect_response provenance chains, which can exceed 200KB for banks with many models.

Response Fields

FieldTypeDetail LevelDescription
idstringmetadataUnique mental model ID
bank_idstringmetadataMemory bank ID
namestringmetadataHuman-readable name
tagslistmetadataTags for filtering
last_refreshed_atstringmetadataWhen the mental model was last updated
created_atstringmetadataWhen the mental model was created
source_querystringcontentThe query used to generate content
contentstringcontentThe generated mental model text
max_tokensintcontentMaximum tokens for the mental model content
triggerobjectcontentTrigger settings (see Automatic Refresh)
reflect_responseobjectfullFull reflect response including based_on provenance facts

Refresh a Mental Model

Re-run the source query to update the mental model with current knowledge:

# Refresh a mental model to update with current knowledge
result = client.refresh_mental_model(
bank_id=BANK_ID,
mental_model_id=mental_model_id
)

print(f"Refresh operation ID: {result.operation_id}")

Refreshing is useful when:

  • New memories have been retained that affect the topic
  • Observations have been updated
  • You want to ensure the mental model reflects current knowledge

Update a Mental Model

Update the mental model's name:

# Update a mental model's metadata
updated = client.update_mental_model(
bank_id=BANK_ID,
mental_model_id=mental_model_id,
name="Updated Team Communication Preferences",
trigger={"refresh_after_consolidation": True} # Enable auto-refresh
)

print(f"Updated name: {updated.name}")

Delete a Mental Model

# Delete a mental model
client.delete_mental_model(
bank_id=BANK_ID,
mental_model_id=mental_model_id
)

Tags and Visibility

Mental models support the same tag system as memories. When you assign tags to a mental model, those tags control both which memories it reads during refresh and when it is surfaced during reflect.

How tags affect mental model refresh

warning

Adding tags to a mental model narrows the pool of source memories its refresh can read from. If no memories carry those tags yet, refresh will return empty content (e.g. "I cannot find any information…") even though direct reflect on the same query works. Backfill tags on the relevant memories first, or override the default via trigger.tags_match / trigger.tag_groups.

When a mental model is refreshed (manually or automatically), it runs an internal reflect call to regenerate its content. If the mental model has tags, that reflect call uses all_strict tag matching — meaning it will only read memories that carry all of the mental model's tags. Untagged memories are excluded.

Mental model tags: ["user:alice"]

During refresh, it reads:
✅ "Alice prefers async communication" — has "user:alice"
✅ "Team uses Slack for announcements" — has "user:alice" (plus other tags)
❌ "Company policy: no meetings on Fridays" — untagged, excluded
❌ "Bob dislikes long meetings" — no "user:alice" tag

This means a mental model tagged ["user:alice"] will also pick up memories tagged ["user:alice", "team"] — extra tags on a memory don't disqualify it. Only the mental model's own tags are required to be present.

How tags affect mental model lookup during reflect

When you call reflect with tags, those same tags are used to filter which mental models the agent can see. A mental model is visible only if its tags overlap with the tags on the reflect request.

For more details on tag matching modes (any, any_strict, all, all_strict) and worked examples, see the Recall tags reference.

Listing mental model tags

GET /v1/default/banks/{bank_id}/tags accepts a source query parameter that selects which tag space to enumerate:

  • source=memories (default) — tags attached to memory units.
  • source=mental_models — tags attached to mental models in this bank.

Use the mental_models source to populate autocomplete or filter UIs over mental-model tags, distinct from the (typically larger) memory tag set.


History

Every time a mental model's content changes (via refresh or manual update), the previous version is saved with a timestamp. You can retrieve the full change log with the history endpoint:

# Get the change history of a mental model
history = client.get_mental_model_history(
bank_id=BANK_ID,
mental_model_id=mental_model_id
)

for entry in history:
print(f"Changed at: {entry['changed_at']}")
print(f"Previous content: {entry['previous_content']}")

Response

The endpoint returns a list of history entries, most recent first:

FieldTypeDescription
previous_contentstring | nullThe content before this change (null if not available)
changed_atstringISO 8601 timestamp of when the change occurred

Each entry captures the content before the change and when it happened. The current content is returned by the standard Get a Mental Model endpoint.

note

History tracking is enabled by default. Set HINDSIGHT_API_ENABLE_MENTAL_MODEL_HISTORY=false to disable it.


Use Cases

Use CaseExample
FAQ AnswersPre-compute answers to common customer questions
Onboarding Summaries"What should new team members know?"
Status Reports"What's the current project status?" refreshed weekly
Policy Summaries"What are our security policies?"

Next Steps

  • Reflect — How the agentic loop uses mental models
  • Observations — How knowledge is consolidated
  • Operations — Track async mental model creation