logo

Work on Your Business, Not the Plumbing.

Work on Your Business, Not the Plumbing.

Your Entire AI Stack in one API.
Portable Memory, State and RAG across 17,000+ LLMs.

We built the world’s fastest, smartest AI memory and paired it with the most flexible stack on the planet. Thousands of LLMs, built-in RAG (or use your own). One API, total control.

user pic

Rob Imbeault, Backboard IO

What’s New with Backboard

A place for releases, results, and community learnings—follow along.

What’s New with Backboard

A place for releases, results, and community learnings—follow along.

Announcement

Feb 12, 2026

Backboard.io Becomes First AI Platform to Lead Both Major Memory Benchmarks

Backboard.io today announced state-of-the-art results across the two leading AI memory benchmarks, LoCoMo and LongMemEval, reinforcing its position as the foundational AI stack for production-grade and agentic systems.

An independent evaluation conducted by NewMathData, a Texas-based engineering firm and AWS Small Partner of the Year, measured Backboard’s performance on the LongMemEval benchmark using the benchmark’s original academic specification. Backboard achieved 93.4% overall accuracy, the highest publicly reported result under consistent methodology and a material margin ahead of other reported systems.

During post-evaluation review, Backboard and the independent evaluator identified multiple instances where Backboard’s responses were marked incorrect despite being more precise and semantically accurate than the benchmark’s expected answer. In these cases, Backboard answered the question as written, incorporating factual context already present in the interaction, while the benchmark’s “gold” answer reflected a narrower or alternate interpretation of the prompt. As a result, the reported LongMemEval score should be considered a conservative lower bound on performance rather than an upper limit.

These results build on Backboard’s previously published 90.1% accuracy on the LoCoMo benchmark, with results publicly available and reproducible via GitHub. Achieving state-of-the-art performance on both benchmarks is uncommon, as most systems optimize for either short-horizon precision or long-horizon persistence, but not both.

Importantly, Backboard did not set out to optimize for benchmarks. The LongMemEval evaluation was initiated and run independently, and the LoCoMo benchmark was explored simply to understand where Backboard fit relative to academic research. The results reflect system-level behavior, not benchmark-specific tuning.

“We didn’t build Backboard to chase benchmarks,” said Rob Imbeault, founder of Backboard.io. “We built it to solve real problems that show up when AI systems run for a long time, across multiple agents, under real constraints. The benchmarks just happened to confirm what we were already seeing in practice.”

Independent Validation of What “Memory” Really Means

In a recent analysis published by the Ottawa Business Journal, Adyasha Maharana, creator of the LoCoMo benchmark and research scientist at Databricks, clarified an important distinction often lost in AI evaluations.

“The dataset is designed to examine not just an LLM but any LLM-based system’s capabilities and blindspots in a fine-grained manner,” Maharana explained. “Raw human performance is somewhere around 88 percent. Breaking the 90-percent threshold requires superhuman consistency in recall and reasoning.”

She further noted that most high-performing frontier models currently score around 80 percent on LoCoMo when evaluated by feeding the full conversation as a single prompt.

“Strictly speaking, this is not memory,” she said. “This is simply understanding whether the LLM pays attention to each part of its input and is able to reason over it correctly. The system built by Backboard.io is a far better attempt at simulating memory as it manifests in humans. It is practical, cheaper, scalable and doesn’t rely solely on brute-force LLM processing for answers.”

This distinction underscores why Backboard’s results reflect more than model capacity. They demonstrate a system-level approach to memory that persists, evolves, and remains reliable over time.

A Complete AI Stack, Not a Bolt-On Component

Backboard.io is not a router, a wrapper, or a memory plugin. It is a unified AI infrastructure stack designed to serve as the starting point for modern AI systems.

From a single API, Backboard provides:

  • Persistent long-term memory


  • Native embeddings and vectorization


  • Retrieval-augmented generation (RAG)


  • Shared memory across agents


  • Access to more than 17,000 large language models, including a bring-your-own API key option.


By integrating memory, embeddings, retrieval, and model access into one system, Backboard eliminates the need for enterprises to stitch together fragile chains of open-source components. Memory is treated as first-class infrastructure, not application logic.

This architecture allows systems to evolve without breaking:

  • Models can be swapped without losing continuity


  • Agents can coordinate while sharing state


  • Retrieval strategies can change without rewrites


  • Systems remain coherent as complexity grows


Making Agentic AI Practical

As interest in agentic AI accelerates, many systems fail to move beyond isolated demos because memory is treated as an afterthought. Without reliable, shared memory, agents fragment, hallucinate, and reset.

Backboard addresses this constraint directly by enabling persistent, shared memory across countless agents, even when those agents operate on different underlying models. When memory is solved, agentic behavior emerges naturally rather than being scripted.

“Agentic AI doesn’t become meaningful because you call something an agent,” said Imbeault. “It becomes meaningful when agents can remember, coordinate, and operate over time. Solving memory is the prerequisite.”

Backboard.io’s architecture is built around Active Temporal Resonance, a memory framework designed to preserve meaning and continuity as interactions unfold. By maintaining temporal coherence rather than reconstructing state through static graphs or repeated retrieval, Backboard enables systems that remain consistent, auditable, and trustworthy at scale.

Built by a Founder Enterprises Already Trust

Imbeault previously founded Assent, a platform trusted by Fortune 100 companies to manage complex supply-chain and regulatory-compliance workflows. That experience informed Backboard’s focus on durability, correctness, and trust from day one.

“Enterprise systems don’t get to reset,” said Imbeault. “If they lose context or trust, they fail. That mindset shaped how we built Backboard.”

What Comes Next

With foundational memory validated across independent and academic benchmarks, Backboard is turning its attention to how teams evaluate and reason about complex AI systems in practice.

The company will soon introduce Switchboard, a new capability designed to help developers and enterprises better understand how different AI system configurations behave under real-world constraints. Additional details will be shared in the coming weeks.

“The future of AI isn’t about clever tricks or bolt-ons,” said Imbeault. “It’s about building systems that can be trusted over time. Memory is the foundation, and that’s where enterprises should start.”

Additional details on Backboard’s benchmark results are available on the company’s website and GitHub repository. The LongMemEval evaluation report and supporting materials will be released publicly.

Announcement

Dec 28, 2025

Backboard.io vs Mem0

When developers compare Backboard to Mem0, the comparison is reasonable if you narrow the lens to one thing only: memory.

Both products aim to help AI systems remember past interactions, user preferences, and relevant context over time. At that level, Mem0 and Backboard are in the same category.

That is where the similarity largely ends.

Comparable on memory. Fundamentally different in scope.

Backboard treats memory as part of a broader system architecture. Mem0 treats memory as the product.

This distinction matters quickly once you move from demos to production.

Memory: surface similarity, different depth

Mem0 provides a lightweight memory layer that developers explicitly read from and write to. It is simple, flexible, and easy to understand. For teams that want to manually control memory operations inside their own orchestration layer, this can be sufficient.

Backboard also provides memory, but with a different design philosophy:

  • Memory reads and writes are handled automatically by default

  • Memory is stateful across full threads, not just isolated snippets

  • Memory accuracy is optimized as a first-class objective, not a side effect

In practice, this means developers spend less time deciding when to store or retrieve memory and more time building product logic. Memory becomes ambient rather than procedural.

If you only compare feature checklists, this difference is easy to miss. If you compare real application behavior over time, it is not.

Accuracy is the real divider

Memory that exists but is inconsistently recalled is worse than no memory at all.

Backboard’s memory system is designed to maximize recall accuracy across long-running, multi-session interactions. This is not just about storing more data. It is about selecting, prioritizing, and retrieving the right information at the right time.

Mem0 focuses on providing a flexible memory abstraction. Backboard focuses on whether the model actually uses the correct memory when it matters.

If you are building anything stateful, agentic, or user-facing, this distinction compounds fast.

Scope: memory product vs AI runtime

Mem0 is a memory layer. It expects you to bring everything else.

Backboard is an AI runtime that includes memory as one component.

Out of the box, Backboard includes:

  • Stateful threads across sessions

  • Long-term and short-term memory management

  • LLM routing across multiple providers

  • Retrieval augmented generation

  • Tool orchestration and custom tools

  • Web search integration

  • Model switching without losing memory

  • Unified billing and observability

This is not about bundling for convenience. It is about reducing architectural risk. Every missing layer you assemble yourself becomes another failure point to maintain.

Developer experience tradeoff

Mem0 gives you control. Backboard gives you leverage.

With Mem0, developers decide when to write memory, when to read memory, and how to merge it into prompts. That can be appealing if you already have a mature orchestration stack.

With Backboard, memory, state, and retrieval are handled by default. You can override behavior when needed, but you are not required to wire every interaction manually.

The practical result is faster time to production and fewer edge cases to debug later.

When Mem0 makes sense

Mem0 is a reasonable choice if:

  • You only need a standalone memory layer

  • You already operate your own routing, tools, and retrieval stack

  • You want full manual control over memory operations

  • Your application scope is narrow and well-defined

When Backboard is the better fit

Backboard is the stronger choice if:

  • Memory accuracy matters in production, not just storage

  • You need state across long-running conversations

  • You want memory to survive model changes

  • You are building agents, assistants, or multi-step workflows

  • You want one system responsible for memory, routing, retrieval, and tools

In short, Backboard is not competing with Mem0 on features alone. It competes by collapsing an entire AI stack into a single coherent runtime where memory actually works.

Bottom line

Yes, Backboard and Mem0 are comparable if you reduce the comparison to memory.

But Backboard is designed for what comes after memory.

If your goal is to store context, Mem0 may be enough.
If your goal is to build reliable, stateful AI systems in production, Backboard solves a much larger problem.

Next step: decide whether you want to assemble your AI stack piece by piece, or whether you want memory, state, and orchestration to work together by default.

One question to pressure-test your decision: are you optimizing for control today, or correctness and velocity six months from now?

Cheers,

Rob

Changelog

Jan 30, 2026

OpenRouter BYOK Now Supported on Backboard

We have launched a new program with OpenRouter that allows OpenRouter users to bring their own keys and run them through Backboard’s stateful API. This unlocks persistent memory, best-in-class RAG, and long-lived agent state without changing how teams already source models.

What’s new

  • Bring Your Own Key (BYOK)
    Use your existing OpenRouter keys directly on Backboard. No new provider contracts, no model lock-in.

  • Stateful API on top of OpenRouter
    Requests gain durable memory, session continuity, and long-running thread state across calls.

  • Integrated RAG and memory
    Pair OpenRouter’s model access with Backboard’s production-grade retrieval, memory persistence, and tooling.

Make everything stateful

Most OpenRouter workflows are stateless by default. That is fine for single prompts but breaks down for agents, assistants, and real applications. By layering Backboard underneath, OpenRouter users get memory and state as first-class primitives, not bolted-on features.

This also avoids a common tradeoff. Teams no longer have to choose between flexible model routing and strong memory architecture. You get both.

Example use case

A small developer team builds a research agent using OpenRouter to test multiple models. With BYOK on Backboard:

  • Conversations persist across sessions

  • Retrieved documents stay grounded over time

  • Agent behavior improves with accumulated memory instead of resetting every call

No changes to model selection. No custom memory system to maintain.

Who this is for

  • Developers already standardized on OpenRouter

  • Teams experimenting with many models but needing consistent state

  • Builders who want memory-native agents without managing infrastructure

Getting started

  1. Connect your OpenRouter key in Backboard

  2. Enable stateful threads and memory

  3. Start building agents that remember

Documentation covers supported models, RAG configuration, and stateful workflows.

Announcement

Feb 12, 2026

Backboard.io Becomes First AI Platform to Lead Both Major Memory Benchmarks

Backboard.io today announced state-of-the-art results across the two leading AI memory benchmarks, LoCoMo and LongMemEval, reinforcing its position as the foundational AI stack for production-grade and agentic systems.

An independent evaluation conducted by NewMathData, a Texas-based engineering firm and AWS Small Partner of the Year, measured Backboard’s performance on the LongMemEval benchmark using the benchmark’s original academic specification. Backboard achieved 93.4% overall accuracy, the highest publicly reported result under consistent methodology and a material margin ahead of other reported systems.

During post-evaluation review, Backboard and the independent evaluator identified multiple instances where Backboard’s responses were marked incorrect despite being more precise and semantically accurate than the benchmark’s expected answer. In these cases, Backboard answered the question as written, incorporating factual context already present in the interaction, while the benchmark’s “gold” answer reflected a narrower or alternate interpretation of the prompt. As a result, the reported LongMemEval score should be considered a conservative lower bound on performance rather than an upper limit.

These results build on Backboard’s previously published 90.1% accuracy on the LoCoMo benchmark, with results publicly available and reproducible via GitHub. Achieving state-of-the-art performance on both benchmarks is uncommon, as most systems optimize for either short-horizon precision or long-horizon persistence, but not both.

Importantly, Backboard did not set out to optimize for benchmarks. The LongMemEval evaluation was initiated and run independently, and the LoCoMo benchmark was explored simply to understand where Backboard fit relative to academic research. The results reflect system-level behavior, not benchmark-specific tuning.

“We didn’t build Backboard to chase benchmarks,” said Rob Imbeault, founder of Backboard.io. “We built it to solve real problems that show up when AI systems run for a long time, across multiple agents, under real constraints. The benchmarks just happened to confirm what we were already seeing in practice.”

Independent Validation of What “Memory” Really Means

In a recent analysis published by the Ottawa Business Journal, Adyasha Maharana, creator of the LoCoMo benchmark and research scientist at Databricks, clarified an important distinction often lost in AI evaluations.

“The dataset is designed to examine not just an LLM but any LLM-based system’s capabilities and blindspots in a fine-grained manner,” Maharana explained. “Raw human performance is somewhere around 88 percent. Breaking the 90-percent threshold requires superhuman consistency in recall and reasoning.”

She further noted that most high-performing frontier models currently score around 80 percent on LoCoMo when evaluated by feeding the full conversation as a single prompt.

“Strictly speaking, this is not memory,” she said. “This is simply understanding whether the LLM pays attention to each part of its input and is able to reason over it correctly. The system built by Backboard.io is a far better attempt at simulating memory as it manifests in humans. It is practical, cheaper, scalable and doesn’t rely solely on brute-force LLM processing for answers.”

This distinction underscores why Backboard’s results reflect more than model capacity. They demonstrate a system-level approach to memory that persists, evolves, and remains reliable over time.

A Complete AI Stack, Not a Bolt-On Component

Backboard.io is not a router, a wrapper, or a memory plugin. It is a unified AI infrastructure stack designed to serve as the starting point for modern AI systems.

From a single API, Backboard provides:

  • Persistent long-term memory


  • Native embeddings and vectorization


  • Retrieval-augmented generation (RAG)


  • Shared memory across agents


  • Access to more than 17,000 large language models, including a bring-your-own API key option.


By integrating memory, embeddings, retrieval, and model access into one system, Backboard eliminates the need for enterprises to stitch together fragile chains of open-source components. Memory is treated as first-class infrastructure, not application logic.

This architecture allows systems to evolve without breaking:

  • Models can be swapped without losing continuity


  • Agents can coordinate while sharing state


  • Retrieval strategies can change without rewrites


  • Systems remain coherent as complexity grows


Making Agentic AI Practical

As interest in agentic AI accelerates, many systems fail to move beyond isolated demos because memory is treated as an afterthought. Without reliable, shared memory, agents fragment, hallucinate, and reset.

Backboard addresses this constraint directly by enabling persistent, shared memory across countless agents, even when those agents operate on different underlying models. When memory is solved, agentic behavior emerges naturally rather than being scripted.

“Agentic AI doesn’t become meaningful because you call something an agent,” said Imbeault. “It becomes meaningful when agents can remember, coordinate, and operate over time. Solving memory is the prerequisite.”

Backboard.io’s architecture is built around Active Temporal Resonance, a memory framework designed to preserve meaning and continuity as interactions unfold. By maintaining temporal coherence rather than reconstructing state through static graphs or repeated retrieval, Backboard enables systems that remain consistent, auditable, and trustworthy at scale.

Built by a Founder Enterprises Already Trust

Imbeault previously founded Assent, a platform trusted by Fortune 100 companies to manage complex supply-chain and regulatory-compliance workflows. That experience informed Backboard’s focus on durability, correctness, and trust from day one.

“Enterprise systems don’t get to reset,” said Imbeault. “If they lose context or trust, they fail. That mindset shaped how we built Backboard.”

What Comes Next

With foundational memory validated across independent and academic benchmarks, Backboard is turning its attention to how teams evaluate and reason about complex AI systems in practice.

The company will soon introduce Switchboard, a new capability designed to help developers and enterprises better understand how different AI system configurations behave under real-world constraints. Additional details will be shared in the coming weeks.

“The future of AI isn’t about clever tricks or bolt-ons,” said Imbeault. “It’s about building systems that can be trusted over time. Memory is the foundation, and that’s where enterprises should start.”

Additional details on Backboard’s benchmark results are available on the company’s website and GitHub repository. The LongMemEval evaluation report and supporting materials will be released publicly.

Announcement

Dec 28, 2025

Backboard.io vs Mem0

When developers compare Backboard to Mem0, the comparison is reasonable if you narrow the lens to one thing only: memory.

Both products aim to help AI systems remember past interactions, user preferences, and relevant context over time. At that level, Mem0 and Backboard are in the same category.

That is where the similarity largely ends.

Comparable on memory. Fundamentally different in scope.

Backboard treats memory as part of a broader system architecture. Mem0 treats memory as the product.

This distinction matters quickly once you move from demos to production.

Memory: surface similarity, different depth

Mem0 provides a lightweight memory layer that developers explicitly read from and write to. It is simple, flexible, and easy to understand. For teams that want to manually control memory operations inside their own orchestration layer, this can be sufficient.

Backboard also provides memory, but with a different design philosophy:

  • Memory reads and writes are handled automatically by default

  • Memory is stateful across full threads, not just isolated snippets

  • Memory accuracy is optimized as a first-class objective, not a side effect

In practice, this means developers spend less time deciding when to store or retrieve memory and more time building product logic. Memory becomes ambient rather than procedural.

If you only compare feature checklists, this difference is easy to miss. If you compare real application behavior over time, it is not.

Accuracy is the real divider

Memory that exists but is inconsistently recalled is worse than no memory at all.

Backboard’s memory system is designed to maximize recall accuracy across long-running, multi-session interactions. This is not just about storing more data. It is about selecting, prioritizing, and retrieving the right information at the right time.

Mem0 focuses on providing a flexible memory abstraction. Backboard focuses on whether the model actually uses the correct memory when it matters.

If you are building anything stateful, agentic, or user-facing, this distinction compounds fast.

Scope: memory product vs AI runtime

Mem0 is a memory layer. It expects you to bring everything else.

Backboard is an AI runtime that includes memory as one component.

Out of the box, Backboard includes:

  • Stateful threads across sessions

  • Long-term and short-term memory management

  • LLM routing across multiple providers

  • Retrieval augmented generation

  • Tool orchestration and custom tools

  • Web search integration

  • Model switching without losing memory

  • Unified billing and observability

This is not about bundling for convenience. It is about reducing architectural risk. Every missing layer you assemble yourself becomes another failure point to maintain.

Developer experience tradeoff

Mem0 gives you control. Backboard gives you leverage.

With Mem0, developers decide when to write memory, when to read memory, and how to merge it into prompts. That can be appealing if you already have a mature orchestration stack.

With Backboard, memory, state, and retrieval are handled by default. You can override behavior when needed, but you are not required to wire every interaction manually.

The practical result is faster time to production and fewer edge cases to debug later.

When Mem0 makes sense

Mem0 is a reasonable choice if:

  • You only need a standalone memory layer

  • You already operate your own routing, tools, and retrieval stack

  • You want full manual control over memory operations

  • Your application scope is narrow and well-defined

When Backboard is the better fit

Backboard is the stronger choice if:

  • Memory accuracy matters in production, not just storage

  • You need state across long-running conversations

  • You want memory to survive model changes

  • You are building agents, assistants, or multi-step workflows

  • You want one system responsible for memory, routing, retrieval, and tools

In short, Backboard is not competing with Mem0 on features alone. It competes by collapsing an entire AI stack into a single coherent runtime where memory actually works.

Bottom line

Yes, Backboard and Mem0 are comparable if you reduce the comparison to memory.

But Backboard is designed for what comes after memory.

If your goal is to store context, Mem0 may be enough.
If your goal is to build reliable, stateful AI systems in production, Backboard solves a much larger problem.

Next step: decide whether you want to assemble your AI stack piece by piece, or whether you want memory, state, and orchestration to work together by default.

One question to pressure-test your decision: are you optimizing for control today, or correctness and velocity six months from now?

Cheers,

Rob

Changelog

Jan 30, 2026

OpenRouter BYOK Now Supported on Backboard

We have launched a new program with OpenRouter that allows OpenRouter users to bring their own keys and run them through Backboard’s stateful API. This unlocks persistent memory, best-in-class RAG, and long-lived agent state without changing how teams already source models.

What’s new

  • Bring Your Own Key (BYOK)
    Use your existing OpenRouter keys directly on Backboard. No new provider contracts, no model lock-in.

  • Stateful API on top of OpenRouter
    Requests gain durable memory, session continuity, and long-running thread state across calls.

  • Integrated RAG and memory
    Pair OpenRouter’s model access with Backboard’s production-grade retrieval, memory persistence, and tooling.

Make everything stateful

Most OpenRouter workflows are stateless by default. That is fine for single prompts but breaks down for agents, assistants, and real applications. By layering Backboard underneath, OpenRouter users get memory and state as first-class primitives, not bolted-on features.

This also avoids a common tradeoff. Teams no longer have to choose between flexible model routing and strong memory architecture. You get both.

Example use case

A small developer team builds a research agent using OpenRouter to test multiple models. With BYOK on Backboard:

  • Conversations persist across sessions

  • Retrieved documents stay grounded over time

  • Agent behavior improves with accumulated memory instead of resetting every call

No changes to model selection. No custom memory system to maintain.

Who this is for

  • Developers already standardized on OpenRouter

  • Teams experimenting with many models but needing consistent state

  • Builders who want memory-native agents without managing infrastructure

Getting started

  1. Connect your OpenRouter key in Backboard

  2. Enable stateful threads and memory

  3. Start building agents that remember

Documentation covers supported models, RAG configuration, and stateful workflows.

Announcement

Feb 12, 2026

Backboard.io Becomes First AI Platform to Lead Both Major Memory Benchmarks

Backboard.io today announced state-of-the-art results across the two leading AI memory benchmarks, LoCoMo and LongMemEval, reinforcing its position as the foundational AI stack for production-grade and agentic systems.

An independent evaluation conducted by NewMathData, a Texas-based engineering firm and AWS Small Partner of the Year, measured Backboard’s performance on the LongMemEval benchmark using the benchmark’s original academic specification. Backboard achieved 93.4% overall accuracy, the highest publicly reported result under consistent methodology and a material margin ahead of other reported systems.

During post-evaluation review, Backboard and the independent evaluator identified multiple instances where Backboard’s responses were marked incorrect despite being more precise and semantically accurate than the benchmark’s expected answer. In these cases, Backboard answered the question as written, incorporating factual context already present in the interaction, while the benchmark’s “gold” answer reflected a narrower or alternate interpretation of the prompt. As a result, the reported LongMemEval score should be considered a conservative lower bound on performance rather than an upper limit.

These results build on Backboard’s previously published 90.1% accuracy on the LoCoMo benchmark, with results publicly available and reproducible via GitHub. Achieving state-of-the-art performance on both benchmarks is uncommon, as most systems optimize for either short-horizon precision or long-horizon persistence, but not both.

Importantly, Backboard did not set out to optimize for benchmarks. The LongMemEval evaluation was initiated and run independently, and the LoCoMo benchmark was explored simply to understand where Backboard fit relative to academic research. The results reflect system-level behavior, not benchmark-specific tuning.

“We didn’t build Backboard to chase benchmarks,” said Rob Imbeault, founder of Backboard.io. “We built it to solve real problems that show up when AI systems run for a long time, across multiple agents, under real constraints. The benchmarks just happened to confirm what we were already seeing in practice.”

Independent Validation of What “Memory” Really Means

In a recent analysis published by the Ottawa Business Journal, Adyasha Maharana, creator of the LoCoMo benchmark and research scientist at Databricks, clarified an important distinction often lost in AI evaluations.

“The dataset is designed to examine not just an LLM but any LLM-based system’s capabilities and blindspots in a fine-grained manner,” Maharana explained. “Raw human performance is somewhere around 88 percent. Breaking the 90-percent threshold requires superhuman consistency in recall and reasoning.”

She further noted that most high-performing frontier models currently score around 80 percent on LoCoMo when evaluated by feeding the full conversation as a single prompt.

“Strictly speaking, this is not memory,” she said. “This is simply understanding whether the LLM pays attention to each part of its input and is able to reason over it correctly. The system built by Backboard.io is a far better attempt at simulating memory as it manifests in humans. It is practical, cheaper, scalable and doesn’t rely solely on brute-force LLM processing for answers.”

This distinction underscores why Backboard’s results reflect more than model capacity. They demonstrate a system-level approach to memory that persists, evolves, and remains reliable over time.

A Complete AI Stack, Not a Bolt-On Component

Backboard.io is not a router, a wrapper, or a memory plugin. It is a unified AI infrastructure stack designed to serve as the starting point for modern AI systems.

From a single API, Backboard provides:

  • Persistent long-term memory


  • Native embeddings and vectorization


  • Retrieval-augmented generation (RAG)


  • Shared memory across agents


  • Access to more than 17,000 large language models, including a bring-your-own API key option.


By integrating memory, embeddings, retrieval, and model access into one system, Backboard eliminates the need for enterprises to stitch together fragile chains of open-source components. Memory is treated as first-class infrastructure, not application logic.

This architecture allows systems to evolve without breaking:

  • Models can be swapped without losing continuity


  • Agents can coordinate while sharing state


  • Retrieval strategies can change without rewrites


  • Systems remain coherent as complexity grows


Making Agentic AI Practical

As interest in agentic AI accelerates, many systems fail to move beyond isolated demos because memory is treated as an afterthought. Without reliable, shared memory, agents fragment, hallucinate, and reset.

Backboard addresses this constraint directly by enabling persistent, shared memory across countless agents, even when those agents operate on different underlying models. When memory is solved, agentic behavior emerges naturally rather than being scripted.

“Agentic AI doesn’t become meaningful because you call something an agent,” said Imbeault. “It becomes meaningful when agents can remember, coordinate, and operate over time. Solving memory is the prerequisite.”

Backboard.io’s architecture is built around Active Temporal Resonance, a memory framework designed to preserve meaning and continuity as interactions unfold. By maintaining temporal coherence rather than reconstructing state through static graphs or repeated retrieval, Backboard enables systems that remain consistent, auditable, and trustworthy at scale.

Built by a Founder Enterprises Already Trust

Imbeault previously founded Assent, a platform trusted by Fortune 100 companies to manage complex supply-chain and regulatory-compliance workflows. That experience informed Backboard’s focus on durability, correctness, and trust from day one.

“Enterprise systems don’t get to reset,” said Imbeault. “If they lose context or trust, they fail. That mindset shaped how we built Backboard.”

What Comes Next

With foundational memory validated across independent and academic benchmarks, Backboard is turning its attention to how teams evaluate and reason about complex AI systems in practice.

The company will soon introduce Switchboard, a new capability designed to help developers and enterprises better understand how different AI system configurations behave under real-world constraints. Additional details will be shared in the coming weeks.

“The future of AI isn’t about clever tricks or bolt-ons,” said Imbeault. “It’s about building systems that can be trusted over time. Memory is the foundation, and that’s where enterprises should start.”

Additional details on Backboard’s benchmark results are available on the company’s website and GitHub repository. The LongMemEval evaluation report and supporting materials will be released publicly.

Announcement

Dec 28, 2025

Backboard.io vs Mem0

When developers compare Backboard to Mem0, the comparison is reasonable if you narrow the lens to one thing only: memory.

Both products aim to help AI systems remember past interactions, user preferences, and relevant context over time. At that level, Mem0 and Backboard are in the same category.

That is where the similarity largely ends.

Comparable on memory. Fundamentally different in scope.

Backboard treats memory as part of a broader system architecture. Mem0 treats memory as the product.

This distinction matters quickly once you move from demos to production.

Memory: surface similarity, different depth

Mem0 provides a lightweight memory layer that developers explicitly read from and write to. It is simple, flexible, and easy to understand. For teams that want to manually control memory operations inside their own orchestration layer, this can be sufficient.

Backboard also provides memory, but with a different design philosophy:

  • Memory reads and writes are handled automatically by default

  • Memory is stateful across full threads, not just isolated snippets

  • Memory accuracy is optimized as a first-class objective, not a side effect

In practice, this means developers spend less time deciding when to store or retrieve memory and more time building product logic. Memory becomes ambient rather than procedural.

If you only compare feature checklists, this difference is easy to miss. If you compare real application behavior over time, it is not.

Accuracy is the real divider

Memory that exists but is inconsistently recalled is worse than no memory at all.

Backboard’s memory system is designed to maximize recall accuracy across long-running, multi-session interactions. This is not just about storing more data. It is about selecting, prioritizing, and retrieving the right information at the right time.

Mem0 focuses on providing a flexible memory abstraction. Backboard focuses on whether the model actually uses the correct memory when it matters.

If you are building anything stateful, agentic, or user-facing, this distinction compounds fast.

Scope: memory product vs AI runtime

Mem0 is a memory layer. It expects you to bring everything else.

Backboard is an AI runtime that includes memory as one component.

Out of the box, Backboard includes:

  • Stateful threads across sessions

  • Long-term and short-term memory management

  • LLM routing across multiple providers

  • Retrieval augmented generation

  • Tool orchestration and custom tools

  • Web search integration

  • Model switching without losing memory

  • Unified billing and observability

This is not about bundling for convenience. It is about reducing architectural risk. Every missing layer you assemble yourself becomes another failure point to maintain.

Developer experience tradeoff

Mem0 gives you control. Backboard gives you leverage.

With Mem0, developers decide when to write memory, when to read memory, and how to merge it into prompts. That can be appealing if you already have a mature orchestration stack.

With Backboard, memory, state, and retrieval are handled by default. You can override behavior when needed, but you are not required to wire every interaction manually.

The practical result is faster time to production and fewer edge cases to debug later.

When Mem0 makes sense

Mem0 is a reasonable choice if:

  • You only need a standalone memory layer

  • You already operate your own routing, tools, and retrieval stack

  • You want full manual control over memory operations

  • Your application scope is narrow and well-defined

When Backboard is the better fit

Backboard is the stronger choice if:

  • Memory accuracy matters in production, not just storage

  • You need state across long-running conversations

  • You want memory to survive model changes

  • You are building agents, assistants, or multi-step workflows

  • You want one system responsible for memory, routing, retrieval, and tools

In short, Backboard is not competing with Mem0 on features alone. It competes by collapsing an entire AI stack into a single coherent runtime where memory actually works.

Bottom line

Yes, Backboard and Mem0 are comparable if you reduce the comparison to memory.

But Backboard is designed for what comes after memory.

If your goal is to store context, Mem0 may be enough.
If your goal is to build reliable, stateful AI systems in production, Backboard solves a much larger problem.

Next step: decide whether you want to assemble your AI stack piece by piece, or whether you want memory, state, and orchestration to work together by default.

One question to pressure-test your decision: are you optimizing for control today, or correctness and velocity six months from now?

Cheers,

Rob

Changelog

Jan 30, 2026

OpenRouter BYOK Now Supported on Backboard

We have launched a new program with OpenRouter that allows OpenRouter users to bring their own keys and run them through Backboard’s stateful API. This unlocks persistent memory, best-in-class RAG, and long-lived agent state without changing how teams already source models.

What’s new

  • Bring Your Own Key (BYOK)
    Use your existing OpenRouter keys directly on Backboard. No new provider contracts, no model lock-in.

  • Stateful API on top of OpenRouter
    Requests gain durable memory, session continuity, and long-running thread state across calls.

  • Integrated RAG and memory
    Pair OpenRouter’s model access with Backboard’s production-grade retrieval, memory persistence, and tooling.

Make everything stateful

Most OpenRouter workflows are stateless by default. That is fine for single prompts but breaks down for agents, assistants, and real applications. By layering Backboard underneath, OpenRouter users get memory and state as first-class primitives, not bolted-on features.

This also avoids a common tradeoff. Teams no longer have to choose between flexible model routing and strong memory architecture. You get both.

Example use case

A small developer team builds a research agent using OpenRouter to test multiple models. With BYOK on Backboard:

  • Conversations persist across sessions

  • Retrieved documents stay grounded over time

  • Agent behavior improves with accumulated memory instead of resetting every call

No changes to model selection. No custom memory system to maintain.

Who this is for

  • Developers already standardized on OpenRouter

  • Teams experimenting with many models but needing consistent state

  • Builders who want memory-native agents without managing infrastructure

Getting started

  1. Connect your OpenRouter key in Backboard

  2. Enable stateful threads and memory

  3. Start building agents that remember

Documentation covers supported models, RAG configuration, and stateful workflows.

Setting a New Standard in AI Memory

The first to lead both industry Memory benchmarks with 90.1% on LoCoMo, and 94.3% on LongMemEval.

Long-Term Conversational Memory (LoCoMo)

Method
Single-Hop (%)
Multi-Hop (%)
Open Domain (%)
Temporal (%)
Overall (%)
Memobase
70.92
46.88
77.17
85.05
75.78
Mem0
67.13
51.15
72.93
55.51
66.88

Backboard.io

89.36

75

91.2

91.9

90.1

Zep
74.11
66.04
67.71
79.79
75.14
LongMem
62.13
47.92
71.12
23.43
58.1
(%)
Single-Hop
Multi-Hop
Open Domain
Temporal
Overall
Mem0
67.13
51.15
72.93
55.51
66.88

89.36

75

91.2

91.9

90.1

Zep
74.11
66.04
67.71
79.79
75.14
Method
Single-Hop (%)
Multi-Hop (%)
Open Domain (%)
Temporal (%)
Overall (%)
Mem0
67.13
51.15
72.93
55.51
66.88

Backboard.io

89.36

75

91.2

91.9

90.1

Zep
74.11
66.04
67.71
79.79
75.14

Long-Term Memory Evaluation (LongMemEval)

Method
Single-Session-Assistant (%)
Single-Session-User (%)
Knowledge Update (%)
Multi-Session (%)
Temporal Reasoning (%)
Single-Session-Preference (%)
Overall (%)
Mastra

(GPT-4o)

82.1
98.6
85.9
79.7
85.7
73.3
84.23

Backboard.io

(GPT-4o)

98.2

97.1

93.6

91.7

91.7

90.0

93.4

Supermemory

(GPT-4o)

96.43
97.14
88.46
71.43
76.69
70.0
81.6
(%)
Single-Session-Assistant
Single-Session-User
Knowledge Update
Multi-Session
Temporal Reasoning
Single-Session-Preference
Overall
Mastra

(GPT-4o)

82.1
98.6
85.9
79.7
85.7
73.3
84.23

(GPT-4o)

98.2

97.1

93.6

91.7

91.7

90.0

93.4

Super-
memory

(GPT-4o)

96.43
97.14
88.46
71.43
76.69
70.0
81.6

Take a Quick Tour

Entire AI Stack in One API

This is just the beginning!

Entire AI Stack in One API

This is just the beginning!

graphic

World's Best Memory

Backboard captures what your AI needs to remember and organizes it for reliable use across any conversation or stack. It stores context, interactions and long-term knowledge, then retrieves the right pieces at the right time. Memory becomes architecture, not a feature.

graphic

World's Best Memory

Backboard captures what your AI needs to remember and organizes it for reliable use across any conversation or stack. It stores context, interactions and long-term knowledge, then retrieves the right pieces at the right time. Memory becomes architecture, not a feature.

Configurable AI Stack

Build exactly the system you want. Combine any model, embedding, vector database and memory setting into a custom stack optimized for your use case. Swap components instantly, tune performance and cost on the fly, and experiment across more than a million possible configurations. It gives developers full control without the overhead of managing complex infrastructure.

Configurable AI Stack

Build exactly the system you want. Combine any model, embedding, vector database and memory setting into a custom stack optimized for your use case. Swap components instantly, tune performance and cost on the fly, and experiment across more than a million possible configurations. It gives developers full control without the overhead of managing complex infrastructure.

Stateful Thread Management

Stateful thread management system supporting all 17,000+ models, allowing on-demand model switching and incorporating robust chunking algorithms that adapt to each model’s context window. Did we mention this is all done for you?

Stateful Thread Management

Stateful thread management system supporting all 17,000+ models, allowing on-demand model switching and incorporating robust chunking algorithms that adapt to each model’s context window. Did we mention this is all done for you?

graphic

RAG Layer

Agentic RAG system with dynamic, scalable capabilities that supports hybrid search (BM25 + vector retrieval), achieving p99 retrieval latency, and can seamlessly expand to accommodate additional documents.

graphic

RAG Layer

Agentic RAG system with dynamic, scalable capabilities that supports hybrid search (BM25 + vector retrieval), achieving p99 retrieval latency, and can seamlessly expand to accommodate additional documents.

Our VC Partners

Our VC Partners

Our VC Partners

Grateful to Mistral (Cohere, Klipfolio), N49P (Spellbook, EvenUP), Garage Capital (Groq, Substack) and Developer Capital (Unified, Glowtify) for believing in the vision!