We’re excited to announce that Rime’s speech models are now available on Together AI, making it easier than ever to build fast, natural, and enterprise-ready AI Voice applications.

    Rime's models are now available on Together AI

    Together AI is the AI Native Cloud purpose-built for AI-native companies training and deploying models at enterprise scale. Built on deep systems research and optimized for GPUs, Together delivers low-latency, cost-efficient inference and customization for always-on production — where throughput, cost, and reliability matter as much as raw speed. With a full suite of compute and software tools supporting over a million developers, Together AI enables teams to run Rime models with best-in-class performance while retaining full ownership and control. 

    This makes Together AI an ideal environment for teams that care deeply about flexibility, performance, and reliability in production.

    By deploying Rime’s models on Together AI, developers can now access state-of-the-art speech generation with significant performance improvements, while being able to run the models within a dedicated environment ensuring data security. 

    Rime’s models are optimized for high-volume enterprise telephony use cases, delivering fast best-in-class latency, concurrency resulting in market-leading call containment and conversion. Rime models outperform all other speech models when comparing ROI for high-volume telephony use-cases.

    Try Rime's models on Together AI

    Why Rime on Together AI

    Built for Speed and Enterprise Scale

    • Low-latency inference designed for real-time and bi-directional streaming applications

    • High-quality, natural-sounding speech generation resulting in higher call containment 

    • Enterprise-ready performance, optimized for reliability and efficiency at scale

    On-Prem Quality, Delivered via the Cloud

    Running Rime models on Together AI’s cloud delivers the quality and control of on-prem deployments, without the operational overhead. Teams get:

    • A dedicated endpoint that’s seamless to deploy

    • Best in-class latency and ability to co-locate models

    • Reliable performance and rigorous benchmarks

    • Enterprise-grade observability in a managed environment

    Spin Up in Seconds

    Anyone can spin up a dedicated endpoint in ~30 seconds. Compute can be co-located with your LLM and speech-to-text stack, dramatically reducing end-to-end latency. This makes Together AI + Rime the best way to build any kind of enterprise telephony voice agent.

    Designed for Voice Agents

    Rime speech models on Together AI offer a massive value proposition for enterprise teams building voice applications

    • Seamless integration with existing Together AI workflows

    • Co-located inference for LLMs, speech-to-text, and text-to-speech

    • Optimized for customer experience and customer support use-cases 

    For new and existing Together AI customers, adding Rime is frictionless. Teams already using Together’s infrastructure, can seamlessly deploy  Rime speech models to their hardware and start shipping voice applications that outperform immediately.

    This makes Rime + Together the fastest path from prototype to enterprise production for voice agents resulting in speed-to-value like never before.


    The platform is built for production AI, with:

    Infrastructure

    • Dedicated GPU capacity with isolated workloads

    • 99.9 percent uptime SLA

    • SOC 2 Type II, HIPAA ready, PCI compliant

    • Global data centers

    • WebSocket streaming support

    • Zero data retention: inputs never stored or used for training

    • Full data ownership and control

    Developer experience

    • Same SDKs and authentication as LLM and STT endpoints

    • Unified pronunciation API across Arcana v2 and Mist v2

    • Single observability and logging surface for entire voice pipeline

    • Model selection and swapping via configuration

    • Professional voice cloning services available

    Get Started

    If you’re already building on Together AI, you can now explore Rime’s speech models and start delivering fast, high-quality voice experiences.

    👉 Check out the Rime models in the Together AI docs and start building today.