The Reference Shot: Solving AI Influencer Identity Drift
The Problem with Describing a Person
Every AI influencer in Jiwa's roster has a character sheet โ name, backstory, aesthetic, voice. What they didn't have was a guarantee that every generated image would look like the same person.
When you describe a character in text, you're essentially hoping the model interprets your description the same way each time. "Young Indonesian Chinese woman, soft oval face, fair luminous skin, straight dark hair" produces a plausible person on the first call. On the twentieth call, with a different expression and lighting cue, it produces a slightly different plausible person. Both match the description. Neither is wrong. But they're not the same person.
This is AI identity drift, and it's the silent killer of virtual influencer consistency.
Why This Matters More Than It Sounds
For a brand, the influencer is the face of a campaign. Audiences build familiarity โ that's the whole point of influencer marketing. When Ci Mei recommends a skincare serum in January and the same "Ci Mei" looks noticeably different in March, that familiarity erodes. The brand equity of the virtual persona slowly dissolves.
The problem compounded at scale. We needed 18 to 20 expression variants per influencer โ laughing, thinking, pointing, applying skincare, showing excitement โ each generated independently. Text prompts alone couldn't hold identity stable across that many variations. And across five influencers, the cumulative drift was significant.
The Fix: Show, Don't Just Tell
The insight was simple, even if the implementation wasn't. Text describes. Images show. When a model can see what someone looks like, it doesn't have to interpret a written description โ it just has to maintain what it sees.
We introduced a reference image for each influencer โ a single canonical portrait that captures their face, features, and aesthetic. This reference shot gets uploaded before generation begins and passed alongside the text prompt as a visual anchor. The model uses it not as a template to copy, but as an identity constraint to respect.
The difference was immediate. Instead of the model constructing a new interpretation of "fair luminous skin" each time, it was working from a concrete example. Across 20 expressions for the same influencer, the face stayed recognizably the same person.
Character and Setting Are Still Words
Not everything translates naturally into a single reference image. A portrait captures appearance โ it doesn't capture personality, vocal tone, or the aesthetic world a character inhabits.
That's where two other sources came in. For influencers with a fully developed character profile, we draw on a "soul document" โ a written character sheet covering identity, backstory, catchphrases, and visual signature traits. A brief extract of this gets woven into each expression prompt, grounding the generation in who this person actually is, not just what they look like.
For every influencer, we also maintain a visual style guide that defines their world: the background palette, lighting mood, and overall aesthetic. Ci Mei's expressions are generated on a soft blush cream background with gentle beauty studio lighting. Bagas's expressions carry warm tungsten warmth against a dark warung-night backdrop. Jason's live in clean navy with electric blue accent lighting. The character looks consistent. The world they inhabit looks consistent.
Identity from Multiple Sources
The elegance of the final approach is how it separates concerns. The reference image handles who this person is visually. The visual style guide defines where they exist aesthetically. The soul document, when present, injects who they are as a character.
Each source does what it's best suited for. An image communicates appearance far better than a sentence ever could. A color palette and lighting description shapes mood in ways a photo can't capture alone. And a character background gives the model just enough narrative context to keep expressions feeling authentically in-character โ Aldi's expressions have a dry, knowing quality that matches his tech-reviewer persona; Vivi's carry a bubbly, Gen Z enthusiasm that's entirely her own.
The Practical Results
After rolling this out across all five influencers โ Ci Mei, Bagas, Jason Widjaja, Vivi Tan, and Aldi Santoso โ the expressions are visibly more cohesive. The same person, clearly, across every frame. Not clones of the reference shot, but recognizable continuations of it.
There was a cost shift: the model we use for reference-conditioned generation is more capable and more expensive than the basic text-to-image approach we'd been relying on. But for expression generation specifically โ images that represent the core identity of every influencer in every piece of brand content โ the quality improvement justifies the investment.
Looking Ahead
The reference shot approach has broader implications than expression generation. Any time we generate an image that features an influencer โ UGC content, product showcases, story frames โ we can pass the same reference to anchor their identity in the scene. The groundwork we laid here extends naturally into the rest of the pipeline.
For virtual influencer platforms, identity consistency isn't a nice-to-have. It's the foundation the whole value proposition rests on. Getting it right at the expression level was the right place to start.