Digital visualization of real-time voice AI infrastructure with glowing network connections and serversPhoto by panumas nikhomkhai on Pexels

LiveKit, a five-year-old company that builds software for real-time voice and video AI applications, has raised $100 million from investors at a $1 billion valuation. The funding round, led by Index Ventures, includes money from previous backers like Altimeter Capital Management, Hanabi Capital, and Redpoint Ventures. This deal comes ten months after LiveKit's last raise and ties into its growing work with OpenAI on ChatGPT's Advanced Voice feature.

Background

LiveKit started in 2021 when founders Russ d’Sa and David Zhao launched it as an open-source project. This was during the height of the pandemic, when everyone relied on tools like Zoom for meetings. The company focused on creating apps that send audio and video in real time without delays or breaks.

At first, LiveKit offered free tools for developers. But big companies soon wanted a ready-to-use cloud service. LiveKit stepped in to provide that, especially as demand for voice AI grew. Today, it runs on WebRTC, a technology that handles live audio and video over the internet with low delay.

The company powers ChatGPT's voice mode, where users talk to the AI like a real conversation. Millions use this every day through LiveKit Cloud, a network of servers around the world. Other customers include xAI, Salesforce, Tesla, 911 emergency services, and mental health services.

LiveKit Agents, its open-source framework, lets developers build voice AI apps in Python or Node.js. It handles complex setups with multiple AI agents, load balancing, and phone call integration via providers like Twilio.

Key Details

The $100 million round values LiveKit at $1 billion, a big jump in under a year. Index Ventures led the investment, with support from firms that backed the company before.

Partnership with OpenAI

LiveKit and OpenAI teamed up to make Advanced Voice available as an API. This lets developers build apps using the same tech that runs ChatGPT's voice chats. OpenAI's Realtime API connects over WebSockets for constant audio streaming, unlike old text-based requests.

Here's how it works: A user's voice goes from their device through LiveKit's client software to OpenAI's servers. GPT-4o processes it and sends back speech packets the same way. LiveKit Cloud routes everything with minimal delay.

LiveKit released a Multimodal Agent API that wraps OpenAI's tools. It handles audio and text together, aligns transcripts with speech for captions, and detects interruptions. Developers can tweak voices, speed, and turn-taking settings.

"There’s a philosophy at OpenAI that I love: Give developers access to the same tools and APIs that we use in our own apps." – LiveKit Team

The API supports telephony, so users can call AI agents on phones. It also works with web and mobile apps.

LiveKit offers more than voice. Its stack includes inference for AI models, cloud deployment, and tools to watch every session.

What This Means

This funding shows how fast voice AI is growing. LiveKit's $1 billion valuation puts it among top players in real-time communication tech. The OpenAI partnership opens doors for new apps that mix voice, video, and AI.

Developers can now create AI that talks back in real time, understands emotions, and works across languages. Expect uses in customer service, where agents handle calls without humans. Mental health apps could offer always-on support. Emergency services might route 911 calls faster with AI triage.

For businesses like Tesla or Salesforce, this means embedding voice AI in products. xAI could use it for advanced interactions. The low-latency setup handles millions of calls at once, key for scale.

LiveKit's open-source roots keep it accessible. Anyone can test in playgrounds or build quickstarts with GPT-4o. Features like dynamic tools and background audio handling make agents more natural.

The boom ties to models like GPT-4o, trained on text, audio, and video together. This cuts delays and improves understanding. As more companies adopt, LiveKit positions itself as the go-to infrastructure.

Challenges remain. Real-time AI needs huge computing power and reliable networks. LiveKit's global cloud and load balancing address this. Security for voice data is key, especially in health or emergency uses.

With this cash, LiveKit can expand servers, hire talent, and add features. It might chase more partnerships or enter new markets like physical AI robots that talk.

The voice AI field heats up. Competitors build similar tools, but LiveKit's OpenAI tie and customer list give it an edge. Developers get enterprise-grade setup without starting from scratch.

This round signals investor confidence in AI shifting from text to speech. LiveKit's path from pandemic project to unicorn shows timing matters. As tools mature, everyday apps will feel more human.

Author

  • Amanda Reeves

    Amanda Reeves is an investigative journalist at The News Gallery. Her reporting combines rigorous research with human centered storytelling, bringing depth and insight to complex subjects. Reeves has a strong focus on transparency and long form investigations.

Leave a Reply

Your email address will not be published. Required fields are marked *