Skip to content

Introduction to Aura Live

Aura Live is a high-performance, real-time transcription (Automatic Speech Recognition) platform built with FastAPI and designed for low-latency live audio processing.

  • Real-time Transcription: Process audio streams and receive text results with minimal latency.
  • Multilingual Support: Support for a growing list of languages and dialects.
  • WebSocket Protocol: A robust, binary-friendly protocol for streaming audio data.
  • REST API: Comprehensive management of languages, models, and sessions.
  • Extensible Architecture: Built to integrate with various ASR engine backends.

The typical lifecycle of an Aura Live interaction follows this flow:

  1. Authentication: The client authenticates using an API token.
  2. Discovery: The client queries the REST API to discover available languages and ASR models.
  3. Session Initiation: The client creates a “Live Session” via a REST endpoint.
  4. Streaming: The client connects to a provided WebSocket URI and streams raw audio data.
  5. Result Retrieval: Aura Live sends transcription results back through the same WebSocket connection.

This documentation is intended for:

  • Developers looking to integrate real-time transcription into their own applications.
  • System Integrators connecting Aura Live to existing audio infrastructure.
  • Users who want to understand the underlying mechanics of the transcription service.

[!NOTE] For administrative tasks (user management, system configuration), please refer to the Admin Dashboard documentation (internal).