So, I picked up vibe coding again in early 2025 once I was making an attempt to learn to make listed chatbots and nice tuned Discord bots that mimic my pal's mannerisms. I found agentic coding when Claude Code was launched and just about turned an addict. It's all I did at night time. Then I bought into brokers, and when ClawBot got here out it was sport over for me (or at the very least my time). So I constructed one and starrt utilizing it to code just about completely, utilizing DIscord to speak with it. I'm looking for a approach out of my present job and I'm hoping this opens up some pathways. Effectively the night/early morning after Valentines Day, once I was lastly in a position to sneak away to my pc and construct, I got here again to a zombified agent and ended up shedding much more progress from the night earlier than than I'd prefer to admit. (Seems once you us discord as your sole technique of communication, exporting your whole chat historical past and even simply telling it to learn again to a sure time-stamp works very well for recovering misplaced reminiscence). In any case, I made a decision to look into methods to enhance its reminiscence, and stumbled throughout some reddit posts and articles that appeared like an excellent place to start out. I swapped my technique from utilizing an ordinary markdown file and storing each 4 hours + on command to a mode of indexing reminiscences with the thought of constructing in a decay system for the reminiscences and a recall and search operate. (Nothing new within the house, nevertheless it was enjoyable to study myself). That's how my first mission was born- Antaris-Reminiscence. It indexes its reminiscences primarily based on precedence, and makes use of native sharded JSONL storage. When it must recall one thing, it makes use of BM25 and decay-weighted looking out, and narrows down the highest 5-10 reminiscences primarily based on the context of the dialog. That was my first module. No RAG, no Vector DB, simply persistent file primarily based reminiscence. Now I'm on V3.0 of antaris-suite, a six Python packages that handles the infrastructure layer of an agent from reminiscence, security, routing, and context utilizing pipeline coordination and shared contracts. Zero exterior dependencies on the core packages. No pulling reminiscences from the cloud, no utilizing different LLMs to kind by means of them, no API keys, nothing. Which, it seems, makes it insanely quick. “`bash If you happen to use OpenClaw: there's a local plugin. **What every package deal really does:** **Antaris-Reminiscence**
**Antaris-Guard**
**Antaris-Router**
-**Antaris-context**
**Antaris Pipeline**
**Antaris-Contract**
— **Benchmarks (Mac Mini M4, 10-core, 32GB):** The Antaris vs mem0 numbers are a direct head-to-head on the identical machine with a dwell OpenAI API key — 50 artificial entries, various corpus sizes (50, 100, 100,000, 500,000, 1,000,000,10 runs averaged. Letta and Zep have been measured individually (totally different methodology — see footnotes). Even with a full pipeline flip of guard + recall + context + routing + ingest antaris measured at 1,000-memory corpus. mem0 determine = measured search p50 (193ms) + measured ingest per entry (312ms). LangChain ConversationBufferMemory: its quick as a result of it's a listing append + recency retrieval — not semantic search. At 1,000+ reminiscences it dumps all the pieces into context. Not equal performance. Zep Cloud measured by way of cloud API from a DigitalOcean droplet (US-West area). Community-inclusive latency. Letta self-hosted: Docker + Ollama (qwen2.5:1.5b + nomic-embed-text) on the identical DigitalOcean droplet. Every ingest generates an embedding by way of Ollama. Not an area in-process comparability. Benchmark scripts are within the repo. For the antaris vs mem0 numbers particularly, you may reproduce them your self in about 60 seconds: “`bash **Engineering choices price noting:** – Storage is apparent JSONL shards + a WAL. Readable, moveable, no lock-in. At 1M entries bulk ingest runs at ~11,600 objects/sec with near-flat scaling (after bulk_ingest repair). — GitHub: Web site: Unique README and the unique thought for the architecure. On the time we consider this to be a novel resolution to the Agent Amnesia drawback, and in addition we've found a variety of these thought have been mentioned earlier than, good quantity of them by no means have, like our Dream State Processing. Pleased to reply questions on structure, the benchmark methodology, or something that appears mistaken. <3 Antaris submitted by /u/fourbeersthepirates |
Subscribe to Updates
Get the latest tech insights from TechnologiesDigest.com on AI, innovation, and the future of digital technology.
Trending
- House Power lays out 15-year house operation plans
- Does Greatest Purchase value match? All the pieces to find out about matching costs on-line and in-store
- Meet OpenMythos: An Open-Supply PyTorch Reconstruction of Claude Mythos The place 770M Parameters Match a 1.3B Transformer
- Why Is The Crypto Market Down As we speak?
- Moonshot AI and Tsinghua Researchers Suggest PrfaaS: A Cross-Datacenter KVCache Structure that Rethinks How LLMs are Served at Scale
- ARPA‑H’s ‘1 Cure’ program bets smarter design can develop most cancers care to extra folks, sooner
- The perfect Kindles in 2026: Skilled really helpful
- How I boosted my transportable photo voltaic panels’ energy by as much as 30% – 11 expert-approved suggestions



![[D] antaris-suite 3.0 (open supply, free) — zero-dependency agent reminiscence, guard, routing, and context administration (benchmarks + 3-model code evaluation inside) [D] antaris-suite 3.0 (open source, free) — zero-dependency agent memory, guard, routing, and context management (benchmarks + 3-model code review inside)](https://technologiesdigest.com/wp-content/uploads/2026/02/D-antaris-suite-30-open-source-free-—-zero-dependency-agent-memory.png)