Long Term AI Memory by creator of Apache Cassandra
cortexdb.ai CortexDB is the long-term memory layer for AI systems — The problem is fundamental: today's AI agents are stateless. Every conversation starts from zero. The dominant approach to giving AI memory — having an LLM rewrite and merge your data on every single write — is lossy, fragile, and ruinously expensive. The LLM decides what to keep and what to throw away, replaces the original with a summary, and that decision is irreversible. Information it deemed unimportant today may be exactly what a future query needs tomorrow. CortexDB takes a fundamentally different approach: every piece of information is appended to an immutable event log and never overwritten. A lightweight LLM extracts entities and relationships asynchronously, but the original data is always preserved — if the ext
cortexdb.ai
CortexDB is the long-term memory layer for AI systems — The problem is fundamental: today's AI agents are stateless. Every conversation starts from zero. The dominant approach to giving AI memory — having an LLM rewrite and merge your data on every single write — is lossy, fragile, and ruinously expensive. The LLM decides what to keep and what to throw away, replaces the original with a summary, and that decision is irreversible. Information it deemed unimportant today may be exactly what a future query needs tomorrow. CortexDB takes a fundamentally different approach: every piece of information is appended to an immutable event log and never overwritten. A lightweight LLM extracts entities and relationships asynchronously, but the original data is always preserved — if the extraction misses something, the raw event is still there for any future query or reprocessing. From this event stream. CortexDB automatically builds a temporal knowledge graph — entities, relationships, causal chains, and provenance — and uses hybrid retrieval combining vector search, full-text matching, graph traversal, and adaptive ranking to assemble the exact context an AI agent needs at query time. The results are not incremental. In controlled benchmarks using identical language models, identical embeddings, and identical test data across five production-scale scenarios, CortexDB achieved a huge gap that is structural, not incidental, because you cannot retrieve information you've already destroyed. The cost difference is equally dramatic because CortexDB's write path uses a lightweight extraction model while rewriting systems burn expensive LLM inference to merge and regenerate entire memory stores on every write operation.
CortexDB scales the same way Cassandra scales — through consistent hashing, partition-aware data placement, and leaderless replication, where every index, every graph shard, and every vector store is scoped to a partition from day one. Adding capacity means adding a node; the cluster rebalances automatically with zero downtime. A single-node deployment is simply a distributed system with one node — the same code path runs whether you have one machine or a hundred. This is not a single-node prototype that will be distributed later. Distribution is the architecture itself, at scale — retrofitting distribution onto a monolithic design costs more than building it right from the start. CortexDB is not a better version of what exists. It is a new layer of infrastructure — the memory layer — built from first principle that scales infinitely unlike any other solution in the market.
Dev.to AI
https://dev.to/prashant_malik_c0d77148e8/long-term-ai-memory-by-creator-of-apache-cassandra-5ap0Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modellanguage modelbenchmark
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
The AI landscape is experiencing unprecedented growth and transformation. This post delves into the key developments shaping the future of artificial intelligence, from massive industry investments to critical safety considerations and integration into core development processes. Key Areas Explored: Record-Breaking Investments: Major tech firms are committing billions to AI infrastructure, signaling a significant acceleration in the field. AI in Software Development: We examine how companies are leveraging AI for code generation and the implications for engineering workflows. Safety and Responsibility: The increasing focus on ethical AI development and protecting vulnerable users, particularly minors. Market Dynamics: How AI is influencing stock performance, cloud computing strategies, and

135,000 OpenClaw Users Just Got a 50x Price Hike. Anthropic Says It's 'Unsustainable.'
Originally published at news.skila.ai A single OpenClaw session can burn through $1,000 to $5,000 in compute. Anthropic was eating that cost on a $200/month Max plan. As of April 4, 2026 at 12pm PT, that arrangement is dead. More than 135,000 OpenClaw instances were running when Anthropic flipped the switch. Claude Pro ($20/month) and Max ($200/month) subscribers can no longer route their flat-rate plans through OpenClaw or any third-party agentic tool. The affected users now face cost increases of up to 50 times what they were paying. This is the biggest pricing disruption in the AI developer tool space since OpenAI killed free API access in 2023. And the ripple effects reach far beyond Anthropic's customer base. What Actually Happened (and Why) Boris Cherny, Head of Claude Code at Anthro

Gemma 4 Complete Guide: Architecture, Models, and Deployment in 2026
Google DeepMind released Gemma 4 on April 3, 2026 under Apache 2.0 — a significant licensing shift from previous Gemma releases that makes it genuinely usable for commercial products without legal ambiguity. This guide covers the full model family, architecture decisions worth understanding, and practical deployment paths across cloud, local, and mobile. The Four Models and When to Use Each Gemma 4 ships in four sizes with meaningfully different architectures: Model Params Active Architecture VRAM (4-bit) Target E2B ~2.3B all Dense + PLE ~2GB Mobile / edge E4B ~4.5B all Dense + PLE ~3.6GB Laptop / tablet 26B A4B 25.2B 3.8B MoE ~16GB Consumer GPU 31B 30.7B all Dense ~18GB Workstation The E2B result is the most surprising: multiple community benchmarks confirm it outperforms Gemma 3 27B on s
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Один промпт заменил мне 3 часа дебага в день
Вечерами, когда большинство уже отдыхает, я зависаю в своём офисе и ковыряюсь с кодом. Тот 14 августа, в 21:45, не был исключением. Я опять сидел над этой задачей, которая съедала по три часа каждый день. Почему это была боль Всё началось с простого: проект на Python, который выглядел как очередное рутинное задание. Однако вычисления упорно выдавали ошибочные результаты. Три дня подряд я безуспешно искал причину. Как обычно, приходилось проверять каждую строчку, каждую переменную. Это было настоящим адом. Для фрилансера с жесткими сроками это катастрофа - теряешь время, не зарабатываешь, а заказчик ждёт. Я собрал промпты по этой теме в PDF. Забери бесплатно: https://t.me/airozov_bot Как я нашёл решение Тогда я решил попробовать ChatGPT, хотя и не особо верил в его чудеса. Вбил проблему в п
ciflow/torchtitan/179381: Update on "[wip][dynamo] Reduce special casing for namedtuple objects"
UserDefinedTupleVariable previously lived in user_defined.py while NamedTupleVariable lived in lists.py and subclassed it across module boundaries. NamedTupleVariable also conflated two unrelated things: Python namedtuples (collections.namedtuple with _tuplegetter descriptors and Type( args) construction) and C-implemented structseqs (torch.return_types. with Type(iterable) construction and tp_new safety checks that reject tuple. new ). Split into three classes, all in user_defined.py: UserDefinedTupleVariable (base): plain tuple subclasses NamedTupleVariable: Python namedtuples, overrides resolve_data_descriptor for _tuplegetter, as_python_constant, as_proxy, reconstruct (uses _make) StructSequenceVariable: torch.return_types.* structseqs, overrides as_python_constant, as_proxy, reconstru



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!