Over the past six months, self-hosted usage of Rime models has grown exponentially. Customers told us they love Rime for its state-of-the-art TTS performance, and they’ve asked us one question again and again:
“Do you do on-prem?”
Today, the answer is an emphatic YES. We’re thrilled to announce that Rime On-Prem is officially out of private beta and available for everyone. See full implementation details on our blog.
Built for Performance, Security, and Scale
Rime On-Prem brings the same high-quality, low-latency speech synthesis our customers rely on: now optimized for your own secure infrastructure. Whether you’re deploying in your private cloud, on edge clusters, or in regulated environments, Rime On-Prem ensures:
🔒 Security and compliance at every layer
⚡ Low latency and real-time response
🔉 Highly accurate pronunciation, tuned for your application
We’ve also partnered with top-tier cloud and edge infrastructure providers to make sure Rime’s voice models are available worldwide and where your users are.
The Biggest Performance Leap Yet for Arcana On-Prem
Alongside the launch, we’re also shipping our biggest performance upgrade ever for Arcana On-Prem, delivering up to 2× higher concurrency on H100 GPUs.
It’s easier than ever to run real-time workloads at scale — securely, efficiently, and exactly where you need them.
Ready to Build at the Edge?
If you’re building high-volume Voice AI applications where security, compliance,latency, and highly accurate pronunciations matter, Rime On-Prem is designed for you.
Email us at: hello@rime.ai

