Our #NeurIPS2025 oral presentation is starting in a few minutes!
Join us:
β°3:30 pm
π Ballroom 6AB
grafting.stanford.edu
arxiv.org/abs/2506.05340
Our #NeurIPS2025 oral presentation is starting in a few minutes!
Join us:
β°3:30 pm
π Ballroom 6AB
grafting.stanford.edu
arxiv.org/abs/2506.05340
Such a great two days of workshops! Fuelled by inspiring talks and excellent reconnections with friends and colleaguesβdefinitely my favorite part of the conference.
Missed my talk on AI Agents: from Language to Multimodal Reasoning?
Summary and slides are here:
www.niebles.net/blog/2025/mm...
Talk is done!
Shared our work on Multimodal AI Agents at the #ICCV2025 Workshop on Multi-Modal Reasoning. π€
All the slides, key papers, and the research journey are consolidated in this new blog post:
πhttps://www.niebles.net/blog/2025/mmagents/
@iccv.bsky.social
We will be presenting Strefer today at Poster 52 9:30-10:30am, join us to learn more about our work on Video-Languange at @salesforce.com AI Research @iccv.bsky.social #ICCV2025
strefer.github.io
arxiv.org/abs/2509.03501
Check out the latest on Strefer: model & data are now available!
arxiv.org/abs/2509.03501
We will see you at #ICCV2025 ποΈ
π’π’ Exciting news!
Our paper, "Exploring Diffusion Transformer Designs via Grafting," has been accepted as an Oral at #NeurIPS2025, with only 77 out of 21k submissions receiving this honor.
πPaper: arxiv.org/abs/2506.05340
πWebsite: grafting.stanford.edu
π§π»βπ»Code: github.com/keshik6/graf...
Strefer: our new work for auto-generating instruction data on spaceβtimeβfocused video tasks: spatiotemporal reasoning, space-time reference understanding, etc. for Video LLMs
β
Auto & scalable
β
Fine-grained, spaceβtimeβgrounded queries
β
Effective
π: arxiv.org/abs/2509.03501
π: strefer.github.io
Check out a new episode of The AI Research Lab - Explained on Multimodal AI.
Had a blast creating this with the @salesforce.com team!
youtu.be/r98jGdLtO6Q
Congrats Chaitanya on winning the BEST PAPER AWARD π₯ π
Check out details of our work:
arxiv.org/abs/2504.12513
Our first #cvpr2025 poster is up!
πCome check it out right now until 13:00
βAdaVid: Adaptive Video-Language Pretrainingβ
πͺ§ExHall D Poster # 203
π arxiv.org/abs/2504.12513
Just finished a day at the #CVPR2025 Area Chair workshop. Lots of interesting discussions and ideas, reconnection with colleagues and friends.
Had the chance to present our ViUnit poster to fellow ACs. If you missed it, come to our Sunday poster session.
See details in the π§΅β¬οΈ
If you're at #CVPR2025, please stop by my posters and say hello! I'd love to chat about our work and all things computer vision. See you in Nashville! π
Last but not least, presenting "ViUniT: Visual Unit Tests for More Robust Visual Programming" #CVPR2025
ποΈ Sun Jun 15, 10:30AM-12:30PM
π ExHall D Poster #346
π Paper: arxiv.org/abs/2412.08859
π Blog: www.niebles.net/blog/2025/vi...
#VisualProgramming #RobustAI
Next, "Re-thinking Temporal Search for Long-Form Video Understanding" #CVPR2025
ποΈ Fri Jun 13, 4PM-6PM
π ExHall D Poster #306
π Paper: arxiv.org/abs/2504.02259
π Website: longvideohaystack.github.io
π» Code: github.com/LongVideoHay...
π Data: huggingface.co/datasets/LVH...
#VideoUnderstanding
I'll also be presenting multiple papers at #CVPR2025! First up: "AdaVid: Adaptive Video-Language Pretraining".
ποΈ Thu Jun 12, 12:00-13:00PM
π ExHall D Poster #202
π Paper: arxiv.org/abs/2504.12513
π Website: chaitanya100100.github.io/AdaVid/
#VideoLanguage #Pretraining
Kicking things off on June 11th by participating in the #CVPR2025 Area Chair workshop! Eager to connect with fellow ACs and colleagues. Let's make this an impactful conference!
Excited to attend #CVPR2025 in Nashville! π€ Looking forward to a fantastic week of cutting-edge computer vision research and connecting with the community.
@cvprconference.bsky.social
Read the full post for more details: "Level up your Agents: Teaching Vision-Language Models to Play by the Rules".
blog: www.niebles.net/blog/2025/vl...
arxiv: arxiv.org/abs/2505.03181
Work with Jake Grigsby, Michael Ryoo and Yuke Zhu
#AI #MachineLearning #DeepLearning
This RL approach effectively aligns VLMs with the demands of interactive decision-making. It's a powerful new pathway for developing more capable and adaptable visual agents using readily available VLM tech.
We tested our approach on PaliGemma, xGen-MM, and MoonDream2 across Gym Cards, BabyAI, and MiniWoB. Results? Substantial improvements in valid action syntax accuracy and task success rates, even starting from noisy data!
This approach works great for offline-to-online fine-tuning, learning from static datasets (even random actions!) and then smoothly transitioning to online learning where the agent gathers new data to refine its policy. Self-improvement is key!
AFSFT helps VLMs overcome challenges like strict action syntax and suboptimal data. It learns from demonstrations and filters out tokens that would lead to invalid syntax or poor choices, even penalizing invalid syntax.
Enter Reinforcement Learning (RL)! Our paper introduces an "offline-to-online" RL technique called Advantage-Filtered Supervised Fine-Tuning (AFSFT) that allows VLMs to learn through trial and error, improving even with imperfect initial data.
Traditional supervised fine-tuning (SFT) has limits β it can't go beyond its training data, and imperfect datasets mean replicating flaws. What if we don't have perfect examples or a good initial VLM?
The catch? VLMs can struggle with the precise rules and structured outputs many agent tasks require, unlike LLMs which excel at function calling and specific syntax. Think describing a button vs. knowing the exact command to click it.
Large Language Models (LLMs) are great for agents, but what happens when we give them "eyes"? VLMs extend this power to process visual info, opening up new possibilities like robotic control and automating tasks by "seeing" your screen.
Just dropped a new blog post: "Level up your Agents: Teaching Vision-Language Models to Play by the Rules"! We're exploring how to make Vision-Language Models (VLMs) even smarter at interactive tasks.
blog: www.niebles.net/blog/2025/vl...
arxiv: arxiv.org/abs/2505.03181
#multimodalAI #agents #VLM
Check out this great intro to Large Action Models, the key engine powering the AI Agent revolution. π€
By @salesforce.com AI Researchβs Shelby Heinecke.
See video here:
youtube.com/watch?v=vlvv...
@salesforce.com #AI Research has a new series called "AI Explained."
π¬ "The AI Research Lab - Explained" debuts with our groundbreaking work on Large Action Models! Sr. Mgr Shelby Heinecke reveals how we're training these specialized models to generate precise, executable actions. t.co/XLhlN2EZyk
Behind every great conference is a team of dedicated reviewers. Congratulations to this yearβs #CVPR2025 Outstanding Reviewers!
cvpr.thecvf.com/Conferences/...