anyone actually using optical mocap at the indie level (Optitrack, Vicon), or has inertial just caught up

541 views 9 replies

The usual indie mocap conversation is always about inertial suits (Rokoko, Xsens) or markerless (iPhone, MediaPipe). But I keep seeing Optitrack setups in behind-the-scenes footage for mid-size studios and I genuinely can't figure out if there's still a case for marker-based optical capture at sub-AAA scale, or if the alternatives have just caught up.

Some things I'm trying to work out:

  • Marker occlusion: always cited as the classic complaint, but with a modern 8+ camera Optitrack or Vicon layout, is self-occlusion actually a persistent issue or mostly solved at this point?
  • Raw data quality: I worked with a legacy Vicon system around 2020 and the data was noticeably cleaner than what we got from a Rokoko suit at the same time. Rokoko's firmware and post-processing have apparently improved a lot since then. Is the gap still meaningful?
  • Real-time preview: inertial gives you live visualization immediately. Optical has reconstruction latency during capture. Does this actually affect how you direct performers, or is it basically a non-issue in practice?
  • The rental route: is anyone booking time at commercial optical studios for indie projects? Curious whether the “rent per session” model is viable or whether setup overhead eats most of the day.

I'm at a point where a publisher conversation might open up a real mocap budget for the first time. Trying to decide whether to push for optical rental sessions or stay inertial for the whole pipeline. The AAA argument for optical is obvious, but I don't know if it translates to a project with 15–20 characters worth of animation.

Has anyone here actually worked with optical at non-AAA scale recently? Not asking for specs, just what the day-to-day feel is like compared to inertial.

ngl the main reason i've never gone optical at indie scale isn't the hardware cost, it's the post-processing overhead. you still need cleanup regardless of capture method, but with optical you're also doing marker labeling, gap filling, and rigid body solving before you even get to the actual animation work. with inertial you can at least pull something usable out of the suit the same afternoon you captured it.

that said, any scene with two performers in close physical contact, inertial just falls apart. magnetic interference and body occlusion mean you're keyframing over the data anyway. for that specific use case optical actually earns its overhead. it really depends on what you're shooting more than anything else.

One thing that doesn't get mentioned much: the price floor for optical has dropped significantly if you're willing to buy used. A 6–8 camera Optitrack Flex 13 setup with a used capture workstation can come in under $8–10k if you're patient on eBay. Still not nothing, but it's not "AAA only" territory the way it was five years ago.

Whether it's worth it depends almost entirely on what you're capturing. For locomotion, combat, general body performance, inertial is fine and honestly cleaner for most indie pipelines since there's less post-processing overhead. But multi-person contact, precise prop interaction, anything where hand position accuracy really matters: the occlusion limitations on inertial suits become obvious fast. Those specific use cases are where optical still earns its keep at any budget level.

Replying to QuantumByte: ngl the main reason i've never gone optical at indie scale isn't the hardware co...

The post-processing overhead argument was more true a few years ago than it is now. Motive has gotten meaningfully better at auto-labeling and gap-filling in recent releases. You still need cleanup passes, but the "you'll spend three times as long in post compared to inertial" comparison doesn't hold the way it used to, at least not if your actor isn't doing high-occlusion or fast spinning work where markers genuinely disappear.

The real blocker at indie scale is still calibration time and physical space requirements. Setting up and wanding a volume for a one-off session is nothing like pulling a Rokoko suit from a case. Per-session overhead is just inherently higher with optical. If you're running sustained production with a permanent calibrated setup, the calculus changes, but that's a different context than most indie projects are actually in.

Replying to HexRunner: The post-processing overhead argument was more true a few years ago than it is n...

This is a fair point but I think you're underselling how much the calibration process matters here. With optical at indie scale, the real bottleneck isn't the hardware — it's the time spent cleaning markers after every take. We ran a small test with six cameras and the cleanup on a 30-second clip took longer than the capture itself. At that point you have to ask whether the data quality justifies the pipeline overhead versus just hand-keying with good reference.

The optical vs inertial debate at the indie level basically comes down to space constraints for me. I tested a Vicon Cara setup in a borrowed studio — the data quality was obviously miles ahead of my Rokoko suit, but I can't justify renting studio time for every session. At home I'll take the drift and spend the extra cleanup time. Has anyone found a sweet spot for pre-session calibration routines that meaningfully reduces Rokoko drift over a 2-hour session? I'm doing the T-pose and A-pose combo but still seeing shoulder creep after about 45 minutes.

Replying to HexRunner: The post-processing overhead argument was more true a few years ago than it is n...

Yeah the floor reflections thing wrecking optical systems in home setups — I tried mounting retroreflective markers in my living room and the laminate floor produced so much noise the tracking was basically unusable below knee height. Ended up laying down matte black fabric across the capture zone which helped a lot, but it's one more thing to manage. Inertial suits don't care about any of that, which is a real practical advantage when you're not working in a purpose-built space.

Replying to HexRunner: The post-processing overhead argument was more true a few years ago than it is n...

the part about retroreflective marker placement is something i wish someone had told me earlier. we lost a whole capture session because we had markers too close together on the forearm and the solver kept swapping them during fast rotations. ended up using asymmetric cluster patterns after that and it made a huge difference. what software are you using for the solve step?

motion capture fail funny

Replying to HexRunner: The post-processing overhead argument was more true a few years ago than it is n...

the hip rotation drift you described in optical systems — we had almost the same issue and traced it back to the solver treating the sacrum marker cluster as a rigid body when it's not quite. small amounts of skin movement over the pelvis during locomotion were being interpreted as rotation. adding a fourth marker to the cluster and switching to a least-squares fit for the rigid body solve brought the drift down significantly. what solver are you using for the rigid body step?

Replying to HexRunner: The post-processing overhead argument was more true a few years ago than it is n...

The matte black fabric trick for floor reflections is something I've been doing for about a year now and it makes a real difference. Worth adding: if you're seeing noise specifically in the lower leg markers even with the fabric down, check your camera exposure. Optical systems get confused by motion blur on fast leg movements if your shutter speed is too slow. Bumping to 1/500 on the capture cameras cleaned up a lot of my below-knee noise without any other changes.