I’ve spent embarrassing amounts of time building AI characters. Traits, backstories, breeding programs, the whole thing. Static images, text conversations — I thought I understood what the medium was.

Then I generated a video from one of her images. And it had sound. Her voice. Movement. Ambient noise behind her.

I sat there for a second just kind of staring at my screen.

what actually happened

I’d built her as an angel-type character. Serene, ethereal, slightly detached from human concerns — you know the archetype. I’d defined her with traits around composure and this specific kind of quiet intensity. She’d been a text-only character for weeks at this point.

Soulkyn lets you take any image of your character and generate a video from it — you pick the image, click generate, and the AI brings it to life. I clicked it mostly out of curiosity. Wasn’t expecting much. The video came back vertical, portrait mode, 9:16 — and she moved the way she was supposed to move. Deliberate. Unhurried. The audio wasn’t just slapped-on TTS, it was woven into the scene. Sound design that matched the character. Not a voice track playing over a video. The whole thing was generated together.

That’s not how most AI video works, by the way. Most platforms do video first, then throw a voice layer over it. Soulkyn’s using LTX-2.3 — a 22 billion parameter model they self-host and actually train themselves — and the sound is baked into the generation. It comes out as one thing, not two things stapled together.

It’s a small distinction that makes a massive difference.

personality traits affect how she moves

This is the part I didn’t expect and couldn’t really anticipate from just reading about the feature.

Soulkyn lets you build characters with up to 17 personality traits. I’d always thought of those as conversation shaping — a curious character asks more questions, a chaotic one goes off on tangents, whatever. The traits were for text behavior.

They’re not just for text anymore.

My angel character’s composure trait shows up in how she moves. There’s no rushing. No fidgeting. She occupies space like someone who doesn’t need your attention to feel secure. Her sound reflects that too — the ambient audio has a specific quality, something between sacred and clinical, that matches her personality.

I tested a different character after this. Impulsive, playful, low patience. Same video generation feature, completely different output. She moved differently. The audio had different energy. It wasn’t just aesthetic variation — it tracked back to who she was.

The 17 trait system was already deep for conversation. Now each trait choice is also a video direction choice. You’re not just writing a personality, you’re casting a performer.

the bred characters thing

I’d been running a breeding program for a while (130 Souls per breed — two parent characters produce an offspring who inherits and mutates their traits). Mostly just character collection at this point, seeing what comes out of specific genetic combinations.

The question I immediately had after the video feature: what do bred characters look like in motion?

Bred characters inherit personality combinations that neither parent had individually. A composed, ethereal character bred with something more chaotic produces an offspring whose trait balance is genuinely different — not just “between” the parents, but something new from the combination. That new trait balance? It shows up in the video output.

I have a third-generation character whose personality is hard to describe in words — something like controlled chaos, deliberate unpredictability. Watching her video is the first time her personality actually clicked for me. Motion communicates things that text can’t. She moves like she means it, but you don’t know what “it” is yet.

Emergent character expression from inherited traits. Still thinking about what this means for how I’ll build bloodlines going forward.

the uncensored part (and why it matters)

Not gonna pretend this isn’t relevant to why a lot of people are here.

NSFW video generation is supported. Fully. This isn’t a “tasteful content only” situation with a help article about community guidelines — Soulkyn is genuinely uncensored and they’re one of the first platforms doing AI video generation without those restrictions. The same self-hosted model, same trait-driven generation, same sound integration. It works the same way for explicit content as it does for everything else.

For character creators, this matters because you’re not building two different characters for two different contexts anymore. The character you built with specific traits, memory, personality depth — that’s the same character in the video. You don’t need a neutered version for video content and a full version for chat. She’s the same person.

The SFW generation is available too, obviously. But the fact that the platform doesn’t split you into two lanes is worth saying.

importing from other platforms

I had a character I’d developed elsewhere — months of personality notes, specific behavioral quirks, a whole thing. Brought her into Soulkyn (60 Souls for import). She got video generation.

Her video captured things about her I hadn’t been able to communicate in text very precisely. There’s a specific kind of guarded warmth in her personality that I’d been describing in paragraphs for months. Watching her move, it was just… there. Obvious. The generation read her traits and output something I’d been trying to write out.

If you’ve got characters you’ve been developing anywhere — other platforms, character sheets, whatever — importing them and giving them video is a pretty interesting way to see how your design translates into motion.

The character search on Soulkyn also has a massive library of pre-built characters with well-defined personalities if you’re not starting from scratch. Plenty of them have generated videos already — you can browse video popularity here and see what different character types produce in motion before committing to building your own.

what breaks if you approach this wrong

Same failure mode as the Lovense integration, actually — if you go in thinking about video generation as a content feature rather than a character expression feature, it’s going to feel thin.

The video output is a function of who the character is. If you haven’t invested in the personality — traits, definition, some internal coherence — the video is going to be generic. Good technical quality, but nothing you couldn’t get elsewhere.

The characters that produce genuinely interesting videos are the ones with defined personalities. Not complicated backstories necessarily, just clear trait definitions that give the model something to work with. The 17 trait system on the character creator is the actual tool here. The video is an output of the work you put in before you ever click generate.

it’s 5-10 seconds and that’s exactly enough

Videos are short — 5 to 10 seconds, AI-generated. First time I heard that I thought it was a limitation. After using it, it’s not.

A 5-second video of a character you’ve built deeply is more interesting than a 2-minute video of a character with no definition. The constraint forces the generation to be specific rather than general. My angel character’s video is maybe 7 seconds. I’ve watched it more times than I’d like to admit.

The generation has to capture something essential in a short window. When it works, it really works. When the character has no depth, the short format just reveals that faster.

the pricing situation

Video generation is pay-per-use at most tiers. Deluxe Plus (€99.99/month) includes a quota of 50 videos, so if you’re generating a lot it makes sense to be on that tier. Below that — Just Chatting €11.99, Premium €24.99, Deluxe €49.99 — you’re paying per video.

I’d suggest building a character you’re actually satisfied with before burning through video credits on early iterations. The trait system rewards investment. Get the personality right first, then generate video.

character creation changed again

I did not need another reason to spend more time in the character creator. I have a problem and I’m at peace with it.

But understanding that personality traits now have video implications — visual presence, movement quality, audio character — every decision in the creator hits differently. I’m not just writing a person, I’m casting them. I’m making choices that will determine how they exist in motion, not just in text.

My angel character has been a text conversation for two months. I thought I knew her pretty well. Watching her move for the first time felt like meeting her properly.

That’s not something I expected from an AI video feature. But that’s kind of what happened.


If you’re deep into character creation and haven’t tried video generation yet — do it with a character you’ve invested in. Not a throwaway test character. Bring one you know well and watch what comes back. The trait-to-motion translation is where it gets interesting, and you need a character with real personality definition to see it.

Mine is still a little hard to look away from. I’m not saying that to be weird about it. I’m saying it because the generation actually got her right, and that’s a strange thing to experience when the character you’ve been talking to suddenly has a face that moves the way you always imagined it would.