You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Your agent remembers, learns from outcomes, and predicts what you need next.
12
12
</p>
13
13
@@ -33,9 +33,17 @@ Most memory layers are glorified vector stores. Store text, retrieve text. Your
33
33
34
34
### Benchmark: LongMemEval
35
35
36
-
Dhee achieves near-perfect retrieval on [LongMemEval](https://arxiv.org/abs/2410.10813), the standard benchmark for long-term conversational memory — temporal reasoning, multi-session aggregation, knowledge updates, and counterfactual tracking across 500+ questions.
36
+
Dhee is being evaluated on [LongMemEval](https://arxiv.org/abs/2410.10813), the standard benchmark for long-term conversational memory — temporal reasoning, multi-session aggregation, knowledge updates, and counterfactual tracking across 500+ questions. Preliminary results are promising.
37
37
38
-
> Evaluation run in progress. Full results and methodology will be published in the benchmark report.
38
+
> Full methodology and results will be published in the benchmark report.
39
+
40
+
---
41
+
42
+
## Status
43
+
44
+
Dhee is **experimental software under active development**. The core 4-operation API (`remember`/`recall`/`context`/`checkpoint`) is stable. Advanced subsystems (belief tracking, policy extraction, episodic indexing) are functional but evolving.
45
+
46
+
Use it. Build on it. But know that internals will change.
39
47
40
48
---
41
49
@@ -93,7 +101,7 @@ Every interface — MCP, Python, CLI, JS — exposes the same 4 operations.
93
101
### `remember(content)`
94
102
Store a fact, preference, or observation.
95
103
96
-
**Hot path**: 0 LLM calls, 1 embedding (~$0.0002). The memory is stored immediately. Echo enrichment (paraphrases, keywords, question-forms that make future recall dramatically better) is deferred to `checkpoint`.
104
+
**Hot path**: 0 LLM calls, 1 embedding (~$0.0002 typical). The memory is stored immediately. Echo enrichment (paraphrases, keywords, question-forms that make future recall dramatically better) is deferred to `checkpoint`.
97
105
98
106
```python
99
107
d.remember("User prefers FastAPI over Flask")
@@ -103,7 +111,7 @@ d.remember("Project uses PostgreSQL 15 with pgvector")
103
111
### `recall(query)`
104
112
Search memory. Returns top-K results ranked by relevance.
105
113
106
-
**Hot path**: 0 LLM calls, 1 embedding (~$0.0002). Pure vector search with echo-boosted re-ranking.
114
+
**Hot path**: 0 LLM calls, 1 embedding (~$0.0002 typical). Pure vector search with echo-boosted re-ranking.
107
115
108
116
```python
109
117
results = d.recall("what database does the project use?")
@@ -161,6 +169,8 @@ d.checkpoint(
161
169
|`checkpoint`| 1 per ~10 memories | 0 |~$0.001 |
162
170
|**Typical session**|**1**|**~15**|**~$0.004**|
163
171
172
+
> Costs assume OpenAI `text-embedding-3-small` at current pricing. Actual costs vary by provider, model, and configuration.
173
+
164
174
---
165
175
166
176
## How It Works (Under the Hood)
@@ -178,7 +188,7 @@ Stores memories in SQLite + a vector index. On the hot path (`remember`/`recall`
178
188
179
189
All of this happens in **1 LLM call per ~10 memories**. Not 4 calls per memory. One batched call.
180
190
181
-
Memory decays naturally (Ebbinghaus curve). Frequently accessed memories get promoted from short-term to long-term. Unused ones fade. ~45% less storage than systems that keep everything forever.
191
+
Memory decays naturally (Ebbinghaus curve). Frequently accessed memories get promoted from short-term to long-term. Unused ones fade. Storage naturally reduces over time as unused memories decay, unlike systems that keep everything indefinitely.
182
192
183
193
### Cognition Engine — Buddhi
184
194
@@ -193,6 +203,17 @@ Zero LLM calls on the hot path. Pure pattern matching + statistics. Persistence
193
203
194
204
Inspired by [Meta's DGM-Hyperagents](https://arxiv.org/abs/2603.19461) — agents that emergently develop persistent memory and performance tracking achieve self-accelerating improvement that transfers across domains. Dhee provides these capabilities as infrastructure.
195
205
206
+
#### Experimental Extensions
207
+
208
+
Beyond the core cognition engine, Dhee includes experimental subsystems that are functional but still evolving:
209
+
210
+
-**Belief store** — confidence-tracked facts with Bayesian updates and contradiction detection
211
+
-**Policy store** — outcome-linked condition→action rules extracted from task completions
212
+
-**Episodic indexing** — structured event extraction for temporal and aggregation queries
213
+
-**Contrastive pairs & heuristic distillation** — learning from what worked vs. what failed
214
+
215
+
These are surfaced through `context()` and `checkpoint()` automatically when enabled.
216
+
196
217
---
197
218
198
219
## Architecture
@@ -245,7 +266,7 @@ m.think("complex question requiring reasoning across memories")
0 commit comments