Michael S. Ryoo
Ph.D.

SUNY Empire Innovation Associate Professor
Department of Computer Science; AI Institute
Stony Brook University


Contact
mryoo-at-cs.stonybrook.edu

I have been with the AI research team at Salesforce. Prior to that, I was with the robotics team at Google DeepMind (and formerly Google Brain) for 5.5 years. I also hold a tenured position in the Department of Computer Science (CS) at Stony Brook University as an associate professor. Previously, I was an assistant professor at Indiana University Bloomington, and was a staff researcher within the Robotics Section of the NASA's Jet Propulsion Laboratory (JPL). I received my Ph.D. from the University of Texas at Austin in 2008 and B.S. from Korea Advanced Institute of Science and Technology (KAIST) in 2004.


Recent News
2025/01
Introducing LLaRA: Supercharging Robot Learning Data for Vision-Language Policy, to appear at ICLR 2025! It is a framework for an efficient transformation of a VLM into a robot Vision-Language-Action (VLA) model.
2024/11
I am organizing CoRL 2025 in Seoul, Korea as a general chair. I look forward to seeing all of you in Seoul!
2024/10
BLIP-3-Video is out! It is an efficient multimodal language model designed for videos. Compared to other models, xGen-MM-Vid represents a video with a fraction of the visual tokens (e.g., 32 vs. 4608 tokens).
2024/05
SARA-RT: Scaling up Robotics Transformers with Self-Adaptive Robust Attention received the Best Paper Award in Robot Manipulation at ICRA 2024.
2023/09
RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control at CoRL 2023!
2023/07
Diffusion Illusions: Hiding Images in Plain Sight received CVPR 2023 Outstanding Demo Award.
It is based on the same lossed used in Peekaboo: Text to Image Diffusion Models are Zero-Shot Segmentors
2023/07
RT-1: Robotics Transformer for Real-World Control at Scale at RSS 2023!
2023/06
Token Turing Machines, a new sequential model modernizing Neural Turing Machines was presented at CVPR 2023. [video]
2022/10
Neural Neural Textures Make Sim2Real Consistent at CoRL 2022
2021/06
Check TokenLearner for images and videos! It learns to adaptively generate a small number of tokens for Transformers, providing better accuracies while also being faster. The paper also appeared at NeurIPS 2021.
2021/04
Neural architecture search for robot reinforcement learning at ICRA 2021: Visionary: Vision Architecture Discovery for Robot Learning

Curriculum Vitae pdf


Publications [by type] [by year]

List of selected publications

Google Scholar: Michael S. Ryoo

Datasets

AViD dataset: Anonymized Videos from Diverse Countries.
MLB-YouTube dataset: an activity recognition dataset with over 42 hours of 2017 MLB post-season baseball videos.
JPL-Interaction dataset: a robot-centric first-person video dataset.
DogCentric Activity dataset: a first-person video dataset taken with dogs.
UT-Interaction dataset: a dataset containing continuous/segmented videos of human-human interactions.

Lab members

Cristina Mata (Stony Brook University CS)
Xiang Li (Stony Brook University CS)
Jongwoo Park (Stony Brook University CS)
Ryan Burgert (Stony Brook University CS)
Kanchana Ranasinghe (Stony Brook University CS)
Abe Leite (Stony Brook University CS)
Yichi Zhang (Stony Brook University CS)
Yoosung Jang (Stony Brook University CS)
E Ro Nguyen (Stony Brook University CS)

Alumni

Kumara Kahatapitiya (PhD 2025; joined Meta)
Jinghuan Shang (PhD 2024; joined Boston Dynamics AI Institute)
Alan Wu (PhD 2023; returned to MIT Lincoln Lab)
Srijan Das (PostDoc 2022; joined UNC Charlotte)
AJ Piergiovanni (PhD 2020; joined Google Brain)

Teaching

CSE378: Intro to Robotics (Fall 2023)
CSE525: Robotics (Spring 2023)
CSE527: Intro to Computer Vision (Fall 2021)
B457/I400: Intro to Computer Vision (Spring 2018)
B659/I590: Vision for Intelligent Robotics (Fall 2017)

Updated 04/2025