You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> 💡 **Temporal reasoning accuracy improved by 159% compared to the OpenAI baseline.**
66
-
67
-
### Details of End-to-End Evaluation on LOCOMO
68
-
69
-
> [!NOTE]
70
-
> Comparison of LLM Judge Scores across five major tasks in the LOCOMO benchmark. Each bar shows the mean evaluation score judged by LLMs for a given method-task pair, with standard deviation as error bars. MemOS-0630 consistently outperforms baseline methods (LangMem, Zep, OpenAI, Mem0) across all task types, especially in multi-hop and temporal reasoning scenarios.
- We use gpt-4o-mini as the processing and judging LLM and bge-m3 as embedding model in MemOS evaluation.
68
+
- The evaluation was conducted under conditions that align various settings as closely as possible. Reproduce the results with our scripts at [`evaluation`](./evaluation).
69
+
- Check the full search and response details at huggingface https://huggingface.co/datasets/MemTensor/MemOS_eval_result.
70
+
> 💡 **MemOS outperforms all other methods (Mem0, Zep, Memobase, SuperMemory et al.) across all benchmarks!**
73
71
74
72
## ✨ Key Features
75
73
@@ -83,6 +81,27 @@ MemOS demonstrates significant improvements over baseline memory solutions in mu
83
81
84
82
## 🚀 Getting Started
85
83
84
+
### ⭐️ MemOS online API
85
+
The easiest way to use MemOS. Equip your agent with memory **in minutes**!
86
+
87
+
Sign up and get started on[`MemOS dashboard`](https://memos-dashboard.openmem.net/cn/quickstart/?source=landing).
88
+
89
+
90
+
### Self-Hosted Server
91
+
1. Get the repository.
92
+
```bash
93
+
git clone https://github.com/MemTensor/MemOS.git
94
+
cd MemOS
95
+
pip install -r ./docker/requirements.txt
96
+
```
97
+
98
+
2. Configure `docker/.env.example` and copy to `MemOS/.env`
Here's a quick example of how to create a **`MemCube`**, load it from a directory, access its memories, and save it.
87
106
88
107
```python
@@ -104,7 +123,7 @@ for item in mem_cube.act_mem.get_all():
104
123
mem_cube.dump("tmp/mem_cube")
105
124
```
106
125
107
-
What about **`MOS`** (Memory Operating System)? It's a higher-level orchestration layer that manages multiple MemCubes and provides a unified API for memory operations. Here's a quick example of how to use MOS:
126
+
**`MOS`** (Memory Operating System) is a higher-level orchestration layer that manages multiple MemCubes and provides a unified API for memory operations. Here's a quick example of how to use MOS:
2. Copy the `configs-example/` directory to a new directory named `configs/`, and modify the configuration files inside it as needed. This directory contains model and API-specific settings.
39
+
## Supported frameworks
40
+
We support `memos-api` and `memos-api-online` in our scripts.
41
+
And give unofficial implementations for the following memory frameworks:`zep`, `mem0`, `memobase`, `supermemory`, `memu`.
23
42
24
43
25
44
## Evaluation Scripts
26
45
27
46
### LoCoMo Evaluation
28
-
⚙️ To evaluate the **LoCoMo** dataset using one of the supported memory frameworks — `memos`, `mem0`, or `zep` — run the following [script](./scripts/run_locomo_eval.sh):
47
+
⚙️ To evaluate the **LoCoMo** dataset using one of the supported memory frameworks — run the following [script](./scripts/run_locomo_eval.sh):
29
48
30
49
```bash
31
50
# Edit the configuration in ./scripts/run_locomo_eval.sh
@@ -45,10 +64,21 @@ First prepare the dataset `longmemeval_s` from https://huggingface.co/datasets/x
45
64
./scripts/run_lme_eval.sh
46
65
```
47
66
48
-
### prefEval Evaluation
67
+
### PrefEval Evaluation
68
+
Downloading benchmark_dataset/filtered_inter_turns.json from https://github.com/amazon-science/PrefEval/blob/main/benchmark_dataset/filtered_inter_turns.json and save it as `./data/prefeval/filtered_inter_turns.json`.
69
+
To evaluate the **Prefeval** dataset — run the following [script](./scripts/run_prefeval_eval.sh):
49
70
50
-
### personaMem Evaluation
71
+
```bash
72
+
# Edit the configuration in ./scripts/run_prefeval_eval.sh
73
+
# Specify the model and memory backend you want to use (e.g., mem0, zep, etc.)
74
+
./scripts/run_prefeval_eval.sh
75
+
```
76
+
77
+
### PersonaMem Evaluation
51
78
get `questions_32k.csv` and `shared_contexts_32k.jsonl` from https://huggingface.co/datasets/bowen-upenn/PersonaMem and save them at `data/personamem/`
52
79
```bash
80
+
# Edit the configuration in ./scripts/run_pm_eval.sh
81
+
# Specify the model and memory backend you want to use (e.g., mem0, zep, etc.)
82
+
# If you want to use MIRIX, edit the the configuration in ./scripts/personamem/config.yaml
0 commit comments