How to Make a VTuber Avatar: Free Tools + Software Compared

Step-by-step guide to making a VTuber avatar with VRoid Studio, Live2D Cubism, or Blender. Free tools, real time estimates, and tracking software tips.

Learning how to make a VTuber model is one of the most searched topics in the VTuber community — and for good reason. Your avatar is your on-screen identity. Knowing how to make a VTuber avatar from scratch, rather than paying a commission artist hundreds of dollars, is a skill that pays off every time you want to refresh your look. This guide walks through three realistic DIY paths: the beginner-friendly VRoid Studio route, the more powerful Live2D Cubism route, and the full-custom Blender pipeline for people who want total control.

By the end you will know which path fits your skill level and time budget, have numbered steps for the fastest route, and understand how to plug your finished model into live tracking software. If you are still at the stage of figuring out whether to become a VTuber at all, check the how to become a VTuber guide first.


TL;DR — Which Path Should You Take?

Your situationBest path
First avatar, want to stream this weekVRoid Studio (Path A)
You draw anime and want 2D live expressionLive2D Cubism (Path B)
Game dev background or willing to spend monthsBlender + UniVRM (Path C)
Budget is zeroVRoid Studio or Blender (both free)
Want the most expressive face trackingLive2D → VTube Studio

Path A — VRoid Studio: The Easiest Way to Make a VTuber Avatar

VRoid Studio is a free desktop app by Pixiv designed specifically as a VTuber model maker. If you want to know the fastest real answer to how to make a VTuber model for free, VRoid is it. You pick a base body, adjust sliders, paint textures, and export a VRM file. No 3D experience needed. Time estimate: 3–8 hours for a first usable avatar.

Download: vroid.com/en/studio

Step 1 — Install VRoid Studio

Download the Windows installer from the official VRoid site. The app is around 600 MB. No registration required to use it. Create a free Pixiv account only if you want to publish to VRoid Hub.

Step 2 — Choose Your Base and Body Proportions

On launch, choose Create New Model and select gender (the label is used for base mesh proportions only — you can adjust everything afterward). Set overall height, head-to-body ratio, and bust/hip sliders. VTubers commonly use stylized proportions with a slightly larger head — aim for a head ratio around 1:5 or 1:6 if you want that anime look.

Step 3 — Design the Face

Click the Face tab. You will find dozens of sliders covering:

  • Eye shape, size, spacing, and iris texture
  • Nose bridge height and width
  • Mouth position and lip thickness
  • Ear shape
  • Skin color and undertone

Each feature has a preset picker plus manual sliders. Spend most of your time here — the face is what viewers see on stream. Change iris texture from the built-in library or import a custom PNG (1024×1024 recommended).

Step 4 — Style the Hair

The Hairstyle tab is where VRoid shines. You add hair groups — top, side, back, bangs, accessories — each as a separate set of guide curves. Drag the control points to shape each strand cluster. Assign color with the built-in gradient editor. More guide points = more detailed hair = higher triangle count. Keep total hair triangles below 20,000 if performance matters.

For a first avatar, use 3–5 hair groups. Complex twin-drill styles with accessories can push you to 10+ groups but the output is still clean.

Step 5 — Dress the Avatar

Open the Outfit tab. Built-in clothing includes shirts, jackets, skirts, pants, and accessories. Each piece has texture variations you can recolor. You can also import a custom texture PNG to replace the base fabric pattern.

For a totally custom outfit, you would export the UV layout, paint in Photoshop or Krita, and reimport. That is a 1–2 hour side task. For a first avatar, the default wardrobe options are enough.

Step 6 — Set Up Default and Blend Shape Expressions

VRoid generates a set of facial expressions automatically: happy, sad, surprised, angry, relaxed, and eye blink. These map directly to tracking software. You can preview them in the Look tab by clicking expression thumbnails.

If you want custom expressions (a wink, a wide grin, a ”>_<” face), open VRM Output Settings → BlendShape. Add new blend shapes there. Note that custom blend shapes beyond the VRM standard set require manual configuration in VTube Studio to assign to hotkeys.

Step 7 — Export as VRM

Go to Export (top menu). Choose Export as VRM. Fill in the author name, license type (choose carefully if you plan to sell), and allowed uses. VRoid auto-generates the rig — you do not rig manually. Click Export, pick a folder, and you have your .vrm file.

Typical file size: 20–80 MB. Triangle count: 30,000–70,000 depending on hair complexity.

Step 8 — Import Into VTube Studio or VSeeFace

  • VTube Studio (Windows + phone camera): Open the desktop app, go to Model → Load Model, select your .vrm file. The app will auto-map standard expressions.
  • VSeeFace (Windows webcam): Drop the .vrm into the VSeeFace models folder, load it from the app.

Both are free. You are now live — your webcam or phone camera drives the avatar’s face in real time.


Path B — Live2D Cubism: More Control, More Work

Live2D Cubism is the industry tool behind virtually every professional-looking 2D VTuber model. The workflow is: draw your character flat in layers (in Photoshop or Clip Studio Paint), import those layers into Cubism, then create a rigged mesh that deforms as the character moves.

Official docs: docs.live2d.com

Cost: Free trial (limited export options). Pro license is around $200–300 or a $2/month subscription via the Live2D site.

Output format: .moc3 + .model3.json bundle. Tracked by VTube Studio.

Time estimate: Several weeks to a few months for your first complete model.

High-Level Steps for Live2D

  1. Draw layered art — separate every part that should move: each eye, each eyebrow, the mouth open/closed, the head, the bangs, the body. Minimum 30–60 layers for a basic model. Save as PSD.
  2. Import PSD into Cubism — layers come in as individual textures.
  3. Warp deformers — place warp and rotation deformers over each part. These are the “bones” of 2D rigging. Assign the head to a rotation deformer so it tilts left/right.
  4. Keyform animation — for each parameter (Head X, Head Y, Eye Open L, etc.) set keyforms at -30, 0, and +30 degrees. Cubism interpolates between them.
  5. Physics — add a physics group for hair and accessories so they swing with head movement.
  6. Export .moc3 — package with the texture atlas and model3.json.
  7. Load in VTube Studio — the .model3.json file is the entry point. VTube Studio maps the standard Live2D parameters to your webcam automatically.

The payoff is expressive, illustrator-quality animation that no 3D tool currently matches for pure anime aesthetics. The cost is real: learning the Cubism parameter system and keyform workflow has a steep initial curve.


Path C — Blender + Unity + UniVRM: Full Custom 3D

This is the path of game developers and 3D artists who want a fully custom mesh with no stylistic constraints. It produces a VRM file like VRoid, but every polygon is yours.

Tools:

  • Blender — free, industry-standard 3D modeling
  • Unity — free Personal tier (required for the UniVRM pipeline)
  • UniVRM — open source Unity package that exports VRM from a Unity scene

Time estimate: 1–3 months minimum for someone new to 3D. Experienced 3D artists can produce a VRM in 2–4 weeks.

High-Level Steps for Blender + UniVRM

  1. Model the character in Blender — polygon budget target: 30,000–60,000 triangles for real-time use.
  2. UV unwrap — lay out UVs cleanly; you will paint or bake textures onto these.
  3. Texture — paint in Substance Painter, Blender’s texture paint mode, or Krita. Standard VRM uses a toon shader, so you want flat color + cel shading, not PBR.
  4. Rig in Blender — create an armature (skeleton) that follows the VRM bone naming convention. UniVRM needs bones named exactly: Hips, Spine, Head, LeftUpperArm, etc.
  5. Weight paint — assign mesh vertices to bones so the body deforms correctly.
  6. Export to FBX → import into Unity.
  7. Install UniVRM in Unity — import via Package Manager using the GitHub URL.
  8. Set up VRM meta — author, thumbnail, license.
  9. Add BlendShapes — create blend shapes for each expression (A/I/U/E/O mouth shapes, blink, joy, angry, sorrow, surprise) in Blender first, then map them in the UniVRM inspector.
  10. Export VRM — builds a .vrm file you can load into VSeeFace or VTube Studio.

How to Make a VTuber Model: DIY Tool Comparison Table

VRoid StudioLive2D CubismBlender + UniVRM
CostFreeFree trial / ~$200–300Free
Output formatVRM (.vrm).moc3VRM (.vrm)
Art styleAnime 3DAnime 2DAnything
Skill requiredNoneIntermediate (drawing + rigging)Advanced (3D modeling)
Time for first avatar3–8 hoursSeveral weeks1–3 months
Expression qualityGoodExcellentGood–Excellent
Tracking softwareVTube Studio, VSeeFaceVTube StudioVTube Studio, VSeeFace
Polygon budget controlLimited (auto-generated)N/A (2D)Full control
Physics (hair/cloth)Built-inBuilt-inBlender → Unity physics
Best forBeginners, fast setup2D artists3D artists, game devs

Adding Facial Expressions and Physics

Regardless of which path you took, two features dramatically improve the avatar’s on-stream presence: extra expressions and physics simulation.

Expressions Beyond the Defaults

VRM models support eight standard blend shapes: Joy, Angry, Sorrow, Fun, A, I, U, E, O, Blink, BlinkLeft, BlinkRight. VTube Studio and VSeeFace map these automatically. To add extras (a wink, a tongue-out, a sweat drop), you need to:

  • In VRoid: define custom blend shapes in the Export Settings panel, then configure them as hotkeys in VTube Studio.
  • In Cubism: add parameter tracks beyond the default set and label them in the model3.json.
  • In Blender/UniVRM: add extra shape keys in Blender and expose them as BlendShapeClips in the UniVRM inspector.

Map your expressions to keyboard shortcuts in your tracking software. Experienced VTubers keep their expression board on a Stream Deck or a secondary keypad.

Physics

Hair, ribbons, loose clothing, and ear accessories all benefit from secondary motion physics. Each tool handles this differently:

  • VRoid: physics groups are configured in the Physics/Collider tab. Add spring groups for each hair group. Adjust stiffness (0–1) and drag.
  • Cubism: physics is a separate editor panel. Define pendulum chains from the root bone outward.
  • UniVRM: spring bones are added as VRMSpringBone components in Unity. Point them at the bone chain you want to swing.

Tuning physics takes experimentation. As a starting point: high stiffness (0.8+) for short hair, low stiffness (0.1–0.3) for long flowing hair or ribbons.


Testing in Tracking Software

Once your model file is ready, testing is the step most beginners skip. Do not skip it.

VTube Studio checklist:

  1. Load the model. Confirm it appears in frame without clipping the floor or ceiling.
  2. Enable face tracking. Open the Face section and verify all expression parameters show movement when you move your face.
  3. Check blink — most camera setups need blink sensitivity adjusted (default is often too high or too low for glasses wearers).
  4. Test mouth sync. Say vowels out loud and confirm the mouth opens/closes.
  5. Test head tilt and rotation up to the extremes of your natural range — look for mesh distortion or clipping at the neck.
  6. Test each custom expression hotkey.
  7. Check the performance overlay: aim for below 10ms render time in the VTube Studio stats panel.

VSeeFace checklist:

The same steps apply. VSeeFace also has a Model Info window showing live blend shape values — useful for diagnosing which parameter is not responding correctly.


Common Errors and How to Fix Them

“Model loads sideways or upside down” This is a coordinate axis mismatch between Blender and Unity. In Blender, apply all transforms (Ctrl+A → All Transforms) before exporting FBX. In Unity, confirm the FBX import Y-up axis is set correctly.

Hair clips through the body Increase the bone collider radius on the neck and shoulders in VRoid’s Physics panel. In UniVRM, add VRMSpringBoneCollider components to the shoulder bones and point the spring bone’s colliders list at them.

Mouth does not open during tracking VTube Studio’s mouth-open parameter needs the camera to see your chin. Adjust webcam angle or lower the mouth-open threshold in the Face Tracking settings panel.

Model looks washed out on stream VRM uses a toon shader expecting sRGB color space. If your streaming software (OBS) is capturing in a different color profile, enable color space correction in OBS’s video source settings.

Expression blend shapes missing in VTube Studio The blend shape names must match VRM standard exactly: Joy, Angry, Blink, A, etc. (case-sensitive). Check your export settings in VRoid or your BlendShapeClip names in UniVRM.

High latency in face tracking VTube Studio’s iOS face tracking (using the phone camera over USB or Wi-Fi) is lower latency than most USB webcams because iPhones use ARKit depth sensors. If you are on webcam and getting sluggish tracking, lower the tracking resolution in VSeeFace’s startup settings.


Voice Strategy Alongside Your VTuber Avatar

Your avatar handles the visual identity. Your voice handles everything else — and for many VTubers it is just as important. A lot of creators want a voice that matches their character persona rather than their everyday voice.

That is where VoxBooster comes in. VoxBooster runs on Windows and processes your microphone in real time — you can shift pitch, apply a neural voice clone, add effects, and run noise suppression all at once. The voice clone feature lets you build a custom voice model from a few minutes of training audio, so your on-screen persona has a consistent voice regardless of your off-camera voice.

VoxBooster works alongside VTube Studio, VSeeFace, and OBS — it outputs through a virtual audio device that any streaming tool picks up as a normal microphone. No setup conflicts with the avatar pipeline. Check the pricing options if you want to try it.

For context on what makes a great avatar once you have the technical side down, the VTuber model guide covers design principles and what to look for when evaluating models. For a full breakdown of voice tools available in 2026, the best voice changer 2026 article covers the evaluation criteria worth knowing.


Frequently Asked Questions

How long does it take to make a VTuber model? With VRoid Studio, a basic avatar takes 3–8 hours. A polished Live2D model takes several weeks. A fully custom Blender+Unity avatar can take months for someone new to 3D.

What is the easiest way to make a VTuber avatar for free? VRoid Studio by Pixiv is the easiest free option. It is a dedicated VTuber model maker with a visual editor, preset parts, and direct VRM export — no 3D experience required.

Can I use my VTuber model without a rigging artist? Yes, if you use VRoid Studio. It auto-rigs the model when you export to VRM format. VTube Studio and VSeeFace both accept VRM files and handle live facial tracking automatically.

How do I make a VTuber avatar move with my face? Export your model as VRM or .moc3, then import it into VTube Studio or VSeeFace. These apps read your facial movements from a webcam or phone camera and map them to the avatar in real time.

What software do VTubers use to track their avatar? VTube Studio is the most popular for 2D Live2D models. VSeeFace is common for VRM 3D models. Both are free. VTube Studio has a paid iOS app for high-quality ARKit face tracking.

Does VRoid Studio have a polygon limit? VRoid exports VRM with no enforced poly cap, but VTube Studio runs best below roughly 70,000 triangles. VRoid’s default export is usually 30,000–60,000 depending on hair complexity.

Can I sell a VTuber avatar I make in VRoid Studio? Yes, under Pixiv’s Terms of Service you can sell VRM models made with VRoid Studio. Read the current Pixiv Commercial Terms before listing, as rules around pre-made assets bundled inside the app may vary.


Conclusion

Making a VTuber avatar is more approachable than it looks from the outside. The fastest path — VRoid Studio — takes a single afternoon and produces a fully tracked, streaming-ready VRM file with no prior 3D experience. The harder paths (Live2D Cubism, Blender + UniVRM) take longer but give you control over every pixel and polygon.

The core of how to make a VTuber work as a content identity is the combination: a recognizable visual (avatar) plus a consistent voice (audio). When you want to know how to make a VTuber model that actually ships — not just gets planned — the answer is to pick one path, finish it, and iterate. Use VRoid to ship your first model this week, and layer in voice processing once the avatar side is solid.

If you are choosing a voice changer to pair with your new model, download VoxBooster and run through the setup — the trial covers everything you need to test the voice clone and effects before committing.

Try VoxBooster — 3-day free trial.

Real-time voice cloning, soundboard, and effects — wherever you already talk.

  • No credit card
  • ~30ms latency
  • Discord · Teams · OBS
Try free for 3 days