Pitch β€” xtan.ai

AI Gesture Control for 3D Workflows ⚑️

πŸ’‘ Idea

MotionCoder is a AI-powered framework that translates gestures and signs in real time into precise CAD/DCC commands β€” for faster 3D workflows, fewer clicks, and better ergonomics.

🚧 The Problem

Today’s 3D workflows (CAD/DCC) are click-heavy, time-consuming, and mode-bound: constantly switching modes, opening dialogs, typing parameters. That costs time, focus, and patience.

βœ… The Solution

MotionCoder recognizes iconic gestures (e.g., line, circle, scissors, handwheel) and turns them in real time into commands with parameters.

Designed primarily for the desktop with a large monitor β€” more comfortable & efficient than VR. VR is optional.

Why now? Modern GPUs with AI acceleration enable low-latency on-prem pipelines with high performance and strong cost-performance β€” often cheaper than ToF setups.

🚨 Gap

Similar ideas for gesture control exist, but they usually lacked end-to-end real-time capability β€” and the integration of sign-language nuances (handshape, fine-tuning, cognitive reference, spatial grammar, coarticulation, timing/rhythm, prosody, etc.). As a result, they saw little adoption or disappeared again. My entrepreneurial experience in sign language, programming, and mechanical engineering closes this gap.

βš™οΈ How it Works

  • Multi-view cameras detect gestures and reconstruct them as 3D gestures.
  • Real-time pipeline: 3D gestures β†’ intent + parameters + continuous values (e.g., angle/Ø/depth).
  • Plugins control the target software (start: Blender, then Unreal, etc.).

🎯 Why It’s Compelling

  • Work faster: Fewer mode switches, more actions per second.
  • Ergonomic: Less mouse micro-work, more focus in the viewport.
  • Intuitive: Iconic gestures are self-explanatory.
  • Accessibility: Signs as a first-class interface ♿️
  • Secure & private: On-prem, no cloud required, data minimization.

πŸ‘₯ Who Benefits?

βš”οΈ Competitive Landscape

  • Direct competition: No solution combines precise multi-view tracking, real-time semantics, and parametric CAD/DCC automation into one end-to-end pipeline.
  • Indirect/partial: Tracking stacks for VR/animation (e.g., MediaPipe, YOLO, Meta Quest) and standard hardware do not deliver industry-grade robustness for parametric CAD/DCC automation in practice.
  • Differentiation: MotionCoder closes the gap: intent + parameters in real time, undo-safe integrations into CAD/DCC.

πŸ‘¨β€πŸ”§ Why Me?

Own CNC workshop; experience in mechanical engineering, drive technology, CAD, kinematics, C/C++ as well as sign-language communication. The idea for MotionCoder grows out of my CNC software development; plus hands-on vision/camera projects for clients.

From practice I know when MotionCoder is faster and more ergonomic than the mouse in everyday work. Gestures are more natural and faster in many scenarios β€” that shop-floor know-how is embedded in xtan.ai.

♻️ Conclusion

Gestures in β†’ result out: fast ⚑️, precise 🎯, performant πŸ’ͺ.

More info: xtan.ai Β· GitHub: xtanai/overview