Tabitha Oanda
Robotics Engineer · Fashion Designer · PhD Researcher at Brown University
Cloth is deformable, slippery, and hard to track — which means off-the-shelf robot setups don't cut it. I build the full stack: a bimanual hardware platform for manipulation, multi-camera perception using foundation vision models to detect and segment fabric across frames in 3D. I use teleoperation tools that make collecting training data practical. Following data collection, I train cloth dynamics models that can be used for model predictive control (MPC) and reinforcement learning policies for complex tasks.
My research is on getting robots to handle fabric reliably. That requires building the whole stack:
- Platform: custom aluminum frame, overhead-mounted bimanual arms, force-sensing silicone grippers
- Perception: multi-camera RGB-D, object detection and segmentation, 3D point tracking
- Data collection: leader-follower teleoperation for imitation learning, custom UMI-inspired data collection gripper
- Learning: cloth dynamics models trained and extended on data from my custom setup
A consistent theme in my research is taking methods developed in academic settings and adapting them to work on real cloth manipulation problems — the kind of contact-rich, deformable-object tasks that matter for industrial textile handling.
Projects
The Sew Unit
A bimanual cloth manipulation platform built from scratch: custom aluminum frame, SO-101 arms, MoveIt planning, and leader-follower teleoperation. All designed, built, and debugged by hand.
View project →Learning Cloth Dynamics
Ran PhysTwin and PGND on cloth data I collected, built the full perception and data pipeline, then explored whether adding visual supervision to dynamics training improves 3D predictions. Results are promising on individual fabrics; active research.
View project →Custom Grippers & Teleop Tools
Designed two custom end-effectors: silicone FSR grippers for contact-aware grasping, and a UMI-inspired handheld teleop gripper with ArUco markers and IMU for data collection.
View project →Policy Training
Training fold policies with ACT, comparing image encoders (ResNet vs DINOv2) and pretrained action models (OpenVLA, Octo, pi0). Baseline results and ongoing experiments.
View project →