Changelog

Nov 13, 2025

AWS Bedrock + Enhanced File Upload

We have expanded our Model Library with full support for AWS Bedrock. Developers can now access a broad set of high performance models through a single Backboard API, with unified memory, configuration, and monitoring.

Anthropic Claude Models

  • Claude 4.5 Haiku
    Fast and efficient with a 200K context window.

  • Claude 4.5 Sonnet
    Strong reasoning performance with prompt caching support.

  • Claude 4 Sonnet and Opus
    Premium models with 200K context windows.

  • Claude 3.x Series
    Includes Sonnet, Haiku, and legacy versions for backward compatibility.

Meta Llama Models

  • Llama 4 Maverick
    One million token context window with tool calling enabled.

  • Llama 4 Scout
    128K context and tool support.

  • Llama 3.3 (70B)
    High performing 70B parameter model.

  • Llama 3.2 Series
    Models ranging from 1B to 90B parameters.

  • Llama 3.1 Series
    8B and 70B instruct tuned models.

Other Providers

  • DeepSeek R1
    Advanced reasoning capabilities for research style workloads.

  • Mistral Pixtral Large
    Multimodal model with strong performance across vision and text tasks.

Enhanced File Upload API

We have added stronger guardrails across assistants, threads, and messages to help developers stay within resource limits without guesswork. Error messages now include detailed usage to make troubleshooting straightforward.

Assistant Level

  • File Limit: 20 files

  • Token Limit: 5,000,000 tokens per file

  • File Size: Maximum 200 MB

Thread Level

  • File Limit: 20 files per thread

  • Combined Limit: 140 total files across assistants, threads, and attachments

Message Attachments

  • Per Message: Up to 10 attachments

  • Per Thread: Up to 100 attachments

  • Token Limit: 1,000,000 tokens per attachment

  • File Size: Maximum 10 MB

Message Input Limits

  • Character Limit: 200,000 characters per message

  • Token Limit: 50,000 tokens per message

These improvements create more predictable development workflows while still giving teams a wide range of flexibility. We will continue pushing these limits upward over time as part of our long term effort to deliver near infinite capacity.

Changelog