Generative AI May 24, 2025

Zoom puts an AI avatar on its earnings call. That makes this a product story

Zoom CEO Eric Yuan opened a quarterly earnings call on May 22 with an AI avatar built with Zoom Clips. Klarna’s CEO had done the same earlier that week. Two public-company CEOs using synthetic video in investor-facing settings makes this a product an...

Zoom puts an AI avatar on its earnings call. That makes this a product story

Zoom’s CEO used an AI avatar on an earnings call. That has real technical consequences

Zoom CEO Eric Yuan opened a quarterly earnings call on May 22 with an AI avatar built with Zoom Clips. Klarna’s CEO had done the same earlier that week. Two public-company CEOs using synthetic video in investor-facing settings makes this a product and governance story, not a novelty demo.

For developers, synthetic video itself isn't the news. We've had that for a while. What's changed is the setting: regulated communication, where latency, identity, auditability, and abuse prevention all matter at the same time.

If a company is willing to put a digital twin in front of analysts and shareholders, it's making a pretty clear statement. It trusts the system enough to use it where reputational damage is expensive and recordings stick around forever.

Why this matters now

Earnings calls are scripted, tightly managed, and legally sensitive. That's exactly why avatar systems fit there.

The format helps. Opening remarks are prepared in advance. The speaker doesn't need to handle live back-and-forth every second. An avatar can deliver the intro, then a live executive can take over for Q&A, where unscripted answers and credibility still carry more weight.

That's probably the most plausible enterprise pattern in the short term:

  • synthetic delivery for prepared remarks
  • human presence for questions, escalation, and anything legally murky

It's practical. No sci-fi required.

It also points to where collaboration software is going. Video platforms spent years polishing live presence. Now they're packaging synthetic presence: generated speech, generated faces, asynchronous clips, translation, and the identity controls needed to keep all of that from becoming a mess.

The stack behind an executive avatar

According to the reference material, Zoom Clips uses an avatar pipeline that combines motion capture, generative face modeling, and neural text-to-speech. It starts with a short calibration step, about 30 seconds of webcam footage plus a voice sample, to train a face model and a Tacotron-style TTS model.

That sounds plausible for a polished enterprise workflow. A webcam-sized talking head doesn't need a Hollywood capture setup. It needs enough source material to model facial geometry, voice characteristics, and timing, plus a rendering pipeline that holds up under real-time or near-real-time constraints.

The rest of the stack matters more than the headline.

Motion and lip sync

The materials describe a hybrid approach using MediaPipe for facial landmarks and GAN-based refinement for lip sync and micro-expressions.

That makes sense. Landmark tracking is fast, mature, and cheap enough for commodity hardware. The refinement layer handles the part users notice immediately: mouth shape, eye movement, and the little facial shifts that keep a face from looking rubbery. Get that wrong and trust falls off a cliff.

Rendering

The reported rendering path uses lightweight WebGL shaders composited over low-latency RTMP or SRT streams.

That suggests Zoom wants avatars to fit inside existing video infrastructure instead of requiring some fragile custom media path. It also keeps deployment sane. Enterprises know how to run systems built around those protocols. They don't want a research project that breaks on ordinary laptops or corporate networks.

Local execution

One of the more interesting details is that avatars reportedly run in a lightweight local client module to reduce latency and avoid sending raw video upstream.

That's smart for two reasons. First, latency. Do some capture, tracking, and rendering on-device and you cut round trips and jitter. Second, privacy. Raw executive footage is sensitive data. Sending less of it to remote inference services is the safer call.

There is a trade-off. Local execution turns the endpoint into part of the threat model. You're not only protecting cloud inference APIs. You're protecting client software, local caches, model artifacts, and signing keys.

The hard problem is identity

A polished avatar demo is easy to admire. The harder engineering work is proving the avatar is authorized.

If a CEO can appear as a synthetic version of themselves on a financial call, every serious deployment needs a clear answer to one question: how do participants know it's legitimate?

The reference material mentions three safeguards:

  • frequency-domain watermarking embedded in the synthetic video
  • OAuth2-based token exchange for avatar authorization
  • audit logs with hash signatures for post-call verification

That's a solid start. It still leaves plenty unresolved.

Watermarks help with provenance, but they degrade under recompression, cropping, or restreaming. They're one layer, not the whole system.

OAuth-style access control fits if the avatar is treated as a principal inside the communications stack. In practice that means a verifiable identity, scoped permissions, session limits, revocation rules, and signed proof tying the avatar back to the human it represents.

Audit logging matters even more in finance, healthcare, and government. If a synthetic executive appears in a high-stakes meeting, you need durable records: who started the session, which model version was used, whether the speech was pre-generated or live-driven, and whether any human operator intervened during the stream.

That's the difference between a neat product feature and something a compliance team might actually sign off on.

Performance will decide whether this spreads

An avatar system used occasionally by a CEO is one thing. A platform used across sales, support, HR, training, and internal communications is another.

Then the questions get familiar:

  • How many concurrent inference sessions can you support?
  • Which parts run on CPU, GPU, or edge accelerators?
  • What happens on a weak laptop with a bad camera and unstable network?
  • How do you degrade gracefully when real-time refinement falls behind?
  • Can you version avatar behavior the way you version application code?

The source material points to containerized inference with Docker and NVIDIA Triton for multilingual support and GPU scaling. That tracks. Any serious enterprise avatar service will need observability, autoscaling, fallback modes, and reproducible model deployments. Prompts, voice assets, policy configs, and model weights all need change control too.

That last piece is easy to underestimate. Once an avatar becomes an official communication channel, model updates turn into governance events.

A lip-sync upgrade is no longer just a quality improvement. It can affect disclosure policy, impersonation risk, accessibility, and legal review.

Where this gets useful quickly

The obvious use case is executive communication across time zones. Pretrained digital twins can handle routine updates while the real person shows up live where judgment matters.

That's useful precisely because it's a little boring.

Customer support and training are the next obvious targets. An avatar can deliver multilingual guided help, keep a consistent visual identity, and generate on demand. Pair that with speech synthesis tuned for local accents and a retrieval layer pulling from current documentation, and it starts to look better than a static FAQ and cheaper than full-time global staffing.

Internal communications may end up being the biggest category. Async Q&A, training updates, policy explainers, and town hall summaries all fit. The source material mentions serverless workflows that can generate avatar responses in under two minutes and route intake through Slack or Teams. That's the sort of integration pattern buyers care about.

There is a line companies shouldn't cross. Using avatars to widen access is fine. Using them to fake executive availability or dodge accountability isn't.

Employees and customers will accept synthetic delivery if it's clearly labeled and used in the right places. They won't love it if leadership starts using it to simulate presence while staying absent.

The standards gap is hard to miss

Right now, each platform is building its own identity and trust model. That won't hold for long.

If Microsoft Teams, Webex, Google Meet, and Zoom all ship enterprise avatar systems, customers will want a common way to verify provenance across platforms. The reference material points to possible W3C or IEEE work on cross-platform identity layers and "avatar passports." The name is a bit much. The need is real.

A few capabilities are likely to become table stakes:

  • signed provenance metadata that survives ordinary distribution
  • explicit disclosure that a participant is synthetic
  • standardized audit events
  • revocation and kill-switch controls
  • policy hooks for regulated environments

Without that, every enterprise ends up stitching together its own trust layer, and a lot of them will be shaky.

What technical leaders should take from this

Zoom's CEO using an avatar on an earnings call doesn't prove digital twins are fully mainstream. It does show the market has moved past toy demos.

If you own collaboration tooling, internal AI platforms, or customer-facing communication systems, the priorities are fairly clear. Start with low-risk, scripted formats. Treat provenance as a product feature from day one. Keep sensitive capture and rendering local where possible. Build operator logs that will hold up under review. Assume every useful avatar feature comes with an abuse case attached.

The companies that get this right won't win on realism alone. They'll win because the systems are fast enough, cheap enough, and trustworthy enough to survive legal review, security scrutiny, and actual users.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI video automation

Automate repetitive creative operations while keeping review and brand control intact.

Related proof
AI video content operations

How content repurposing time dropped by 54%.

Related article
Character.AI adds AvatarFX video generation, Scenes, and social Streams

Character.AI built its audience on text chat with synthetic personalities. Now it wants those characters to move, talk, and circulate in a social feed. The latest rollout adds three pieces at once: AvatarFX for short AI-generated videos, Scenes for s...

Related article
Kaltura buys eSelf for $27M to add real-time conversational avatars

Kaltura is paying $27 million for eSelf.ai, an Israeli startup that builds real-time conversational avatars. By big tech standards, that's a small deal. For enterprise software, it still matters. Kaltura already has a sizable video business, with rou...

Related article
Character.AI introduces AvatarFX, a video model for animating chatbot avatars

Character.AI has unveiled AvatarFX, a video generation model built to animate chatbot characters from either text prompts or still images. It's in closed beta for now. The pitch is simple: take a static avatar, give it a script, and render a speaking...