Changelog
This changelog highlights user-facing changes only. Internal maintenance, CI/CD, and infrastructure updates are omitted.
For full release details, see GitHub Releases.
0.4.17
Features
- Added a manual retry option for failed asynchronous operations. (
dcaacbe4) - You can now change/update tags on an existing document. (
1b4ad7f4) - Added history tracking and a diff view for mental model changes. (
e2baca8b) - Added observation history tracking with a UI diff view to review changes over time. (
576473b6) - File uploads can now choose a parser per request, with configurable fallback chains. (
99220d05) - Added an extension hook that runs after file-to-Markdown conversion completes. (
1d17dea2)
Improvements
- Operations view now supports filtering by operation type and has more reliable auto-refresh behavior. (
f7a60f89) - Added token limits for “source facts” used during consolidation and recall to better control context usage. (
5d05962d) - Improved bank selector usability by truncating very long bank names in the dropdown. (
1e40cd22)
Bug Fixes
- Fixed webhook schema issues affecting multi-tenant retain webhooks. (
32a4882a) - Fixed file ingestion failures by stripping null bytes from parsed file content before retaining. (
cd3a6a22) - Fixed tool selection handling for OpenAI-compatible providers when using named tool_choice. (
1cdfb7c2) - Improved consolidation behavior to prioritize a bank’s mission over an ephemeral-state heuristic. (
00ccf0b2) - Fixed database migrations to correctly handle mental model embedding dimension changes. (
7accac94) - Fixed file upload failures caused by an Iris parser httpx read timeout. (
fa3501d4) - Improved reliability of running migrations by serializing Alembic upgrades within the process. (
f88b50a4) - Fixed Google Cloud Storage authentication when using Workload Identity Federation credentials. (
d2504ac5) - Fixed the bank selector to refresh the bank list when the dropdown is opened. (
0ad8c2d0)
0.4.16
Features
- Added Webhooks with
consolidation.completedandretain.completedevents. (abbf874d)
Improvements
- Improved OpenClaw recall/retention controls. (
d425e93c) - Improved search/reranking quality by switching combined scoring to multiplicative boosts. (
aa8e5475) - Improved performance of observation recall by 40x on large banks. (
ad2cf72a) - Improved server shutdown behavior by capping graceful shutdown time and allowing a forced kill on a second Ctrl+C. (
4c058b4b)
Bug Fixes
- Fixed an async deadlock risk by running database schema migrations in a background thread during startup. (
e0a2ac63) - Fixed webhook delivery/outbox processing so transactions don’t silently roll back due to using the wrong database schema name. (
75b95106) - Fixed observation results to correctly resolve and return related chunks using source_memory_ids. (
cb6d1c46) - Fixed MCP bank-level tool filtering compatibility with FastMCP 3.x. (
f17406fd) - Fixed crashes when an LLM returns invalid JSON across all retries (now handled cleanly instead of raising a TypeError). (
66423b85) - Fixed observations without source dates to preserve missing (None) temporal fields instead of incorrectly populating them. (
891c33b1)
0.4.15
Features
- Added observation_scopes to control the granularity/visibility of observations. (
55af4681) - List documents API now supports filtering by tags (and fixes the q parameter description). (
1d70abfe) - Added PydanticAI integration for persistent agent memory. (
cab5a40f) - Added richer entity label support (optional labels, free-form values, multi-value fields, and UI polish). (
9b96becc) - Added support for timestamp="unset" so content can be retained without a date. (
f903948a) - OpenClaw can now automatically retain the last n+2 turns every n turns (default n=10). (
ad1660b3) - Added configurable Gemini/Vertex AI safety settings for LLM calls. (
73ef99e7) - Added extension hooks to customize root routing and error headers. (
e407f4bc)
Improvements
- Improved recall performance by fetching all recall chunks in a single query. (
61bf428b) - Improved recall/retain performance and scalability for large memory banks. (
7942f181)
Bug Fixes
- Fixed the TypeScript SDK to send null (not undefined) when includeEntities is false. (
15f4b876) - Prevented reflect from failing with context_length_exceeded on large memory banks. (
77defd96) - Fixed a consolidation deadlock caused by retrying after zombie processing tasks. (
c2876490) - Fixed observations count in the control plane that always showed 0. (
eaeaa1f2) - Fixed ZeroEntropy rerank endpoint URL and ensured the MCP retain async_processing parameter is handled correctly. (
f6f1a7d8) - Fixed JSON serialization issues and logging-related exception propagation when using the claude_code LLM provider. (
ecb833f4) - Added bank-scoped request validation to prevent cross-bank/invalid bank operations. (
5270aa5a)
0.4.14
Features
- Add Chat SDK integration to give chatbots persistent memory. (
fed987f9) - Allow configuring which MCP tools are exposed per memory bank, and expand the MCP tool set with additional tools and parameters. (
3ffec650) - Enable the bank configuration API by default. (
4d030707) - Support filtering graph-based memory retrieval by tags. (
0bb5ca4c) - Add batch observations consolidation to process multiple observations more efficiently. (
0aa7c2b3) - Add OpenClaw options to toggle autoRecall and exclude specific providers. (
3f9eb27c) - Add a ZeroEntropy reranker provider option. (
17259675)
Improvements
- Increase customization options for reflect, retain, and consolidation behavior. (
2a322732) - Include source document metadata in fact extraction results. (
87219b73)
Bug Fixes
- Raise a clear error when embedding dimensions exceed pgvector HNSW limits (instead of failing later at runtime). (
8cd65b98) - Fix multi-tenant schema isolation issues in storage and the bank config API. (
b180b3ad) - Ensure LiteLLM embedding calls use the correct float encoding format to prevent embedding failures. (
58f2de70) - Improve recall performance by reducing memory usage during retrieval. (
9f0c031d) - Handle observation regeneration correctly when underlying memories are deleted. (
ac9a94ad) - Fix reflect retrieval to correctly populate dependencies and enforce full hierarchical retrieval. (
8b1a4658) - Fix OpenClaw health checks by passing the auth token to the health endpoint. (
40b02645)
0.4.13
Features
- Switched the default OpenAI LLM to gpt-4o-mini. (
325b5cc1) - Observation recall now includes the source facts behind recalled observations. (
5569d4ad) - Added CrewAI integration to enable persistent memory. (
41db2960)
Bug Fixes
- Fixed npx hindsight-control-plane failing to run. (
0758827d) - Improved MCP compatibility by aligning the local MCP implementation with the server and removing the deprecated stateless parameter. (
ea8163c5) - Fixed Docker startup failures when using named Docker volumes. (
ac739487) - Prevented reranker crashes when an upstream provider returns an error. (
58c4d657) - Improved accuracy of fact temporal ordering by reducing per-fact time offsets. (
c3ef1555) - Client timeout settings are now properly respected. (
dcaa9f14) - Fixed documents not being tracked when fact extraction returns zero facts. (
f78278ea)
0.4.12
Features
- Accept and ingest PDFs, images, and common Office documents as inputs. (
224b7b74) - Add the Iris file parser for improved document parsing support. (
7eafba66) - Add async Retain support via provider Batch APIs (e.g., OpenAI and Groq) for higher-throughput ingestion. (
40d42c58) - Allow Recall to return chunks only (no memories) by setting max_tokens=0. (
7dad9da0) - Add a Go client SDK for the Hindsight API. (
2a47389f) - Add support for the pgvectorscale (DiskANN) vector index backend. (
95c42204) - Add support for Azure pg_diskann vector indexing. (
476726c2)
Improvements
- Improve reliability of async batch Retain when ingesting large payloads. (
aefb3fcf) - Improve AI SDK tooling to make it easier to work with Hindsight programmatically. (
d06a0259)
Bug Fixes
- Ensure document tags are preserved when using the async Retain flow. (
b4b5c44a) - Fix OpenClaw ingestion failures for very large content (E2BIG). (
6bad6673) - Harden OpenClaw behavior (safer shell usage, better HTTP mode handling, and more reliable initialization), including per-user banks support. (
c4610130) - Improve Python client async API consistency and reduce connection drop issues via keepalive timeout fixes. (
8114ef44)
0.4.11
Features
- Added support for LiteLLM SDK as an embeddings and reranking provider. (
e408b7e) - Expanded Postgres search support with additional text/vector extensions, including TimescaleDB pg_textsearch and vchord/pgvector options. (
d871c30) - Added hierarchical configuration scopes (system, tenant, bank) for more flexible multi-tenant setup and overrides. (
8d731f2) - Added reverse proxy/base-path support for running Hindsight behind a proxy. (
93ddd41) - Added MCP tools to create, read, update, and delete mental models. (
f641b30) - Added a "docs" skill for agents/tools to access documentation-oriented capabilities. (
dd1e098) - Added an OpenClaw configuration option to skip recall/retain for specific providers. (
fb7be3e)
Improvements
- Improved LiteLLM gateway model configuration for more reliable provider/model selection. (
7d95a00) - Exposed actual LLM token usage in retain results to improve cost/usage visibility. (
83ca669) - Added user-initiated attribution to request context to improve async task and usage attribution. (
90be7c6) - Added OpenTelemetry tracing for improved request traceability and observability. (
69dec8e) - Helm chart: split TEI embedding and reranker into separate deployments for independent scaling and rollout. (
43f9a8b) - Helm chart: added PodDisruptionBudgets and per-component affinity controls for more resilient scheduling. (
9943957)
Bug Fixes
- Fixed a recursion issue in memory retention that could cause failures or runaway memory usage. (
4f11210) - Fixed Reflect API serialization/schema issues for "based_on" so reflections are returned and stored correctly. (
f9a8a8e) - Improved MCP server compatibility by allowing extra tool arguments when appropriate and fixing bank ID resolution priority. (
7ee229b) - Added missing trust_code environment configuration support. (
60574ee) - Hardened the MCP server with fixes to routing/validation and more accurate usage metering. (
e798979) - Fixed the slim Docker image to include tiktoken to prevent runtime tokenization errors. (
6eec83b) - Fixed MCP operations not being tracked correctly for usage metering. (
888b50d) - Helm chart: fixed GKE deployments overriding the configured HINDSIGHT_API_PORT. (
03f47e2)
0.4.10
Features
- Provided a slimmer Docker distribution to reduce image size and speed up pulls. (
f648178) - Added Markdown support in Reflect and Mental Models content. (
c4ef090) - Added built-in Supabase tenant extension for running Hindsight with Supabase-backed multi-tenancy. (
e99ee0f) - Added TenantExtension authentication support to the MCP endpoint. (
fedfb49)
Improvements
- Improved MCP tool availability/routing based on the endpoint being used. (
d90588b)
Bug Fixes
- Stopped logging database usernames and passwords to prevent credential leaks in logs. (
c568094) - Fixed OpenClaw sessions wiping memory on each new session. (
981cf60) - Fixed hindsight-embed profiles not loading correctly. (
0430588) - Fixed tagged directives so they correctly apply to tagged mental models. (
278718d) - Fixed a cast error that could cause failures at runtime. (
093ecff)
Other
- Added a docker-compose example to simplify local deployment and testing. (
5179d5f)
0.4.9
Features
- New AI SDK integration. (
7e339e1) - Add a Python SDK for running Hindsight in embedded mode (HindsightEmbedded). (
d3302c9) - Add streaming support to the hindsight-litellm wrappers. (
665877b) - Add OpenClaw support for connecting to an external Hindsight API and using dynamic per-channel memory banks. (
6b34692)
Improvements
- Improve the mental models experience in the control plane UI. (
7097716) - Reduce noisy Hugging Face logging output. (
34d9188)
Bug Fixes
- Improve recall endpoint reliability by handling timeouts correctly and rejecting overly long queries. (
dd621a6) - Improve /reflect behavior with Claude Code and Codex providers. (
a43d208) - Fix OpenClaw shell argument escaping for more reliable command execution. (
63e2964)
0.4.8
Features
- Added profile support for
hindsight-embed, enabling separate embedding configurations/workspaces. (6c7f057) - Added support for additional LLM backends, including OpenAI Codex and Claude Code. (
539190b)
Improvements
- Enhanced OpenClaw and
hindsight-embedparameter/config options for easier configuration and better defaults. (749478d) - Added OpenClaw plugin configuration options to select LLM provider and model. (
8564135) - Server now prints its version during startup to simplify debugging and support requests. (
1499ce5) - Improved tracing/debuggability by propagating request context through asynchronous background tasks. (
44d9125) - Added stronger validation and context for mental model create/refresh operations to prevent invalid requests. (
35127d5)
Bug Fixes
- Improved embedding CLI experience with richer logs and isolated profiles to avoid cross-contamination between runs. (
794a743) - Operation validation now runs correctly in the worker process, preventing invalid background operations from slipping through. (
96f0e54) - Fixed unreliable behavior when using a custom PostgreSQL schema. (
3825506)
0.4.7
Features
- Add extension hooks to validate and customize mental model operations. (
9c3fda7) - Add support for using an external embedding API provider in OpenClaw plugin (with additional OpenClaw compatibility fixes). (
4b57b82)
Improvements
- Speed up container startup by preloading the tiktoken encoding during Docker image builds. (
039944c)
Bug Fixes
- Prevent PostgreSQL insert failures by stripping null bytes from text fields before saving. (
ef9d3a1) - Fix worker schema selection so it uses the correct default database schema. (
d788a55) - Honor an already-set HINDSIGHT_API_DATABASE_URL instead of overwriting it in the hindsight-embed workflow. (
f0cb192)
0.4.6
Improvements
- Improved OpenClaw configuration setup to make embedding integration easier to configure. (
27498f9)
Bug Fixes
- Fixed OpenClaw embedding version binding/versioning to prevent mismatches when using the embed integration. (
1163b1f)
0.4.5
Bug Fixes
- Fixed occasional failures when retaining memories asynchronously with timestamps. (
cbb8fc6)
0.4.4
Bug Fixes
- Fixed async “retain” operations failing when a timestamp is provided. (
35f0984) - Corrected the OpenClaw daemon integration name to “openclaw” (previously “openclawd”). (
b364bc3)
0.4.3
Features
- Add Vertex AI as a supported LLM provider. (
c2ac7d0,49ae55a) - Add Bearer token authentication for MCP and propagate tenant authentication across MCP requests. (
0da77ce)
Improvements
- CLI: add a --wait flag for consolidate and a --date filter for listing documents. (
ff20bf9)
Bug Fixes
- Fix worker polling deadlocks to prevent background processing from stalling. (
f4f86e3) - Improve reliability of Docker builds by retrying ML model downloads. (
ecc590c) - Fix tenant authentication handling for internal background tasks and ensure the control-plane forwards required auth to the dataplane. (
03bf13e) - Ensure tenant database migrations run at startup and workers use the correct tenant schema context. (
657fe02) - Fix control-plane graph endpoint errors when upstream data is missing. (
751f99a)
Other
- Rename the default bot/user identity from "moltbot" to "openclaw". (
728ce13)
0.4.2
Features
- Added Clawdbot/Moltbot/OpenClaw integration. (
12e9a3d)
Improvements
- Added additional configuration options to control LLM retry behavior. (
3f211f0) - Added real-time logs showing a detailed timing breakdown during consolidation runs. (
8781c9f)
Bug Fixes
- Fixed hindsight-embed crashing on macOS. (
c16ccc2)
0.4.1
Features
- Added support for using a non-default PostgreSQL schema by default. (
2b72e1f)
Improvements
- Improved memory consolidation performance (benchmarking and optimizations). (
b43ef98)
Bug Fixes
- Fixed the /version endpoint returning an incorrect version. (
cfcc23c) - Fixed mental model search failing due to UUID type mismatch after text-ID migration. (
94cc0a1) - Added safer PyTorch device detection to prevent crashes on some environments. (
67c4788) - Fixed Python packages exposing an incorrect version value. (
fccbdfe)
0.4.0
Observations, Mental Models, new Agentic Reflect and Directives, read the announcement.
Features
- Added support for providing a custom prompt for memory extraction. (
3172e99) - Expanded the LiteLLM integration with async retain/reflect support, cleaner API, and support for tags/mission (including passing API keys correctly). (
1d4879a) - Added a new worker service to run background tasks at scale. (
4c79240) - MCP retain now supports timestamps. (
b378f68) - Added support for installing skills via
npx add-skill. (ec22317)
Improvements
- CLI retain-files now accepts more file types. (
1eeced3)
Bug Fixes
- Fixed a macOS crash in the embed daemon caused by an XPC connection issue. (
e5fc6ee) - Fixed occasional extraction in the wrong language. (
87d4a36) - Fixed PyTorch model initialization issues that could cause startup failures (meta tensor/init problems). (
ddaa5f5)
Features
- Add memory tags so you can label and filter memories during recall/reflect. (
20c8f8b) - Allow choosing different AI providers/models per operation. (
e6709d5) - Add Cohere support for embeddings and reranking. (
4de0730) - Add configurable embedding dimensions and OpenAI embeddings support. (
70de23e) - Support custom base URLs for OpenAI-style embeddings and Cohere endpoints. (
fa53917) - Add LiteLLM gateway support for routing LLM/embedding requests. (
d47c8a2) - Add multilingual content support to improve handling and retrieval across languages. (
c65c6a9) - Add delete memory bank capability. (
4b82d2d) - Add backup/restore tooling for memory banks. (
67b273d)
Improvements
- Add retention modes to control how memories are extracted and stored. (
fb31a35) - Add offline (optional) database migrations to support restricted/air-gapped deployments. (
233bd2e) - Add database connection configuration options for more flexible deployments. (
33fac2c) - Load .env automatically on startup to simplify configuration. (
c06d9b4) - Expose an operation ID from retain requests so async/background processing can be tracked. (
1dacd0e) - Add per-request LLM token usage metrics for monitoring and cost tracking. (
29a542d) - Add LLM call latency metrics for performance monitoring. (
5e1f13e) - Include tenant in metrics labels for better multi-tenant observability. (
1ffc2a4) - Add async processing option to MCP retain tool for background retention workflows. (
37fc7fb)
Bug Fixes
- Fix extension loading in multi-worker deployments so all workers load extensions correctly. (
f5f3fca) - Improve recall performance by batching recall queries. (
5991308) - Improve retrieval quality and stability for large memory banks (graph/MPFP retrieval fixes). (
6232e69) - Fix entities list being limited to 100 entities. (
26bf571) - Fix UI only showing the first 1000 memories. (
67c1a42) - Fix duplicated causal relationships and improve token usage during processing. (
49e233c) - Improve causal link detection accuracy. (
2a00df0) - Make retain max completion tokens configurable to prevent truncation issues. (
7715a51) - Fix Python SDK not sending the Authorization header, preventing authenticated requests. (
39e3f7c) - Fix stats endpoint missing tenant authentication in multi-tenant setups. (
d6ff191) - Fix embedding dimension handling for tenant schemas in multi-tenant databases. (
6fe9314) - Fix Groq free-tier compatibility so requests work correctly. (
d899d18) - Fix security vulnerability (qs / CVE-2025-15284). (
b3becb6) - Restore MCP tools for listing and creating memory banks. (
9fd5679)
0.2.0
Features
- Add additional model provider support, including Anthropic Claude and LM Studio. (
787ed60) - Add multi-bank access and new MCP tools for interacting with multiple memory banks via MCP. (
6b5f593) - Allow supplying custom entities when retaining memories via the retain endpoint. (
dd59bc8) - Enhance the /reflect endpoint with max_tokens control and optional structured output responses. (
d49e820)
Improvements
- Improve local LLM support for reasoning-capable models and streamline Docker startup for local deployments. (
eea0f27) - Support operation validator extensions and return proper HTTP errors when validation fails. (
ce45d30) - Add configurable observation thresholds to control when observations are created/updated. (
54e2df0) - Improve graph visualization to the control plane for exploring memory relationships. (
1a62069)
Bug Fixes
- Fix MCP server lifecycle handling so MCP lifespan is correctly tied to the FastAPI app lifespan. (
6b78f7d)
0.1.15
Features
- Add the ability to delete documents from the web UI. (
f7ff32d)
Improvements
- Improve the API health check endpoint and update the generated client APIs/types accordingly. (
e06a612)
0.1.14
Bug Fixes
- Fixes the embedded “get-skill” installer so installing skills works correctly. (
0b352d1)
0.1.13
Improvements
- Improve reliability by surfacing task handler failures so retries can occur when processing fails. (
904ea4d) - Revamp the hindsight-embed component architecture, including a new daemon/client model and CLI updates for embedding workflows. (
e6511e7)
Bug Fixes
- Fix memory retention so timestamps are correctly taken into account. (
234d426)
0.1.12
Features
- Added an extensions system for plugging in new operations/skills (including built-in tenant support). (
2a0c490) - Introduced the hindsight-embed tool and a native agentic skill for embedding/agent workflows. (
da44a5e)
Improvements
- Improved reliability when parsing LLM JSON by retrying on parse errors and adding clearer diagnostics. (
a831a7b)
Bug Fixes
- Fixed structured-output support for Ollama-based LLM providers. (
32bca12) - Adjusted LLM validation to cap max completion tokens at 100 to prevent validation failures. (
b94b5cf)
0.1.11
Bug Fixes
- Fixed the standalone Docker image and control plane standalone build process so standalone deployments build correctly. (
2948cb6)
0.1.10
This release contains internal maintenance and infrastructure changes only.
0.1.9
Features
- Simplified local MCP installation and added a standalone UI option for easier setup. (
1c6acc3)
Bug Fixes
- Fixed the standalone Docker image so it builds and starts reliably. (
b52eb90) - Improved Docker runtime reliability by adding required system utilities (procps). (
ae80876)
0.1.8
Bug Fixes
- Fix bank list responses when a bank has no name. (
04f01ab) - Fix failures when retaining memories asynchronously. (
63f5138) - Fix a race condition in the bank selector when switching banks. (
e468a4e)
0.1.7
This release contains internal maintenance and infrastructure changes only.
0.1.6
Features
- Added support for the Gemini 3 Pro and GPT-5.2 models. (
bb1f9cb) - Added a local MCP server option for running/connecting to Hindsight via MCP without a separate remote service. (
7dd6853)
Improvements
- Updated the Postgres/pg0 dependency to a newer 0.11.x series for improved compatibility and stability. (
47be07f)
0.1.5
Features
- Added LiteLLM integration so Hindsight can capture and manage memories from LiteLLM-based LLM calls. (
dfccbf2) - Added an optional graph-based retriever (MPFP) to improve recall by leveraging relationships between memories. (
7445cef)
Improvements
- Switched the embedded Postgres layer to pg0-embedded for a smoother local/standalone experience. (
94c2b85)
Bug Fixes
- Fixed repeated retries on 400 errors from the LLM, preventing unnecessary request loops and failures. (
70983f5) - Fixed recall trace visualization in the control plane so search/recall debugging displays correctly. (
922164e) - Fixed the CLI installer to make installation more reliable. (
158a6aa) - Updated Next.js to patch security vulnerabilities (CVE-2025-55184, CVE-2025-55183). (
f018cc5)
0.1.3
Improvements
- Improved CLI and UI branding/polish, including new banner/logo assets and updated interface styling. (
fa554b8)
0.1.2
Bug Fixes
- Fixed the standalone Docker image so it builds/runs correctly. (
1056a20)