Change the repository type filter
All
Repositories list
28 repositories
- Official code for NeurIPS 2025 paper "GRIT: Teaching MLLMs to Think with Images"
- Official implementation of the NeurIPS 2025 paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"
EvoPresent
PublicMMWorld
PublicOfficial repo of the ICLR 2025 paper "MMWorld: Towards Multi-discipline Multi-faceted World Model Evaluation in Videos"SafeKey
Public[EMNLP 2025] Official code for the paper "SafeKey: Amplifying Aha-Moment Insights for Safety Reasoning"MSSBench
Public[ICLR 2025] Official codebase for the ICLR 2025 paper "Multimodal Situational Safety"Mojito
PublicOfficial repo for the paper "Mojito: Motion Trajectory and Intensity Control for Video Generation""iReason
PublicMLRM-Halu
PublicVLMbench
PublicNeurIPS 2022 Paper "VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation"MiniGPT-5
PublicOfficial implementation of paper "MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens"EditRoom
PublicMMIR
PublicProbMed
Public[ACL 2025 Findings] "Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"- Codebase of ACL 2023 Findings "Aerial Vision-and-Dialog Navigation"
llm_coordination
PublicCode repository for the NAACL 2025 paper "LLM-Coordination: Evaluating and Analyzing Multi-agent Coordination Abilities in Large Language Models"swap-anything
PublicOfficial implementation of the ECCV paper "SwapAnything: Enabling Arbitrary Object Swapping in Personalized Visual Editing"ComCLIP
PublicOfficial implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"Screen-Point-and-Read
PublicCode repo for "Read Anywhere Pointed: Layout-aware GUI Screen Reading with Tree-of-Lens Grounding"R2H
PublicOfficial implementation of the EMNLP 2023 paper "R2H: Building Multimodal Navigation Helpers that Respond to Help Requests"Discffusion
PublicOfficial repo for the TMLR paper "Discffusion: Discriminative Diffusion Models as Few-shot Vision and Language Learners"Naivgation-as-wish
PublicOfficial implementation of the NAACL 2024 paper "Navigation as Attackers Wish? Towards Building Robust Embodied Agents under Federated Learning"T2IAT
PublicPEViT
PublicOfficial implementation of AAAI 2023 paper "Parameter-efficient Model Adaptation for Vision Transformers"- Code for the EMNLP 2021 Oral paper "Are Gender-Neutral Queries Really Gender-Neutral? Mitigating Gender Bias in Image Search" https://arxiv.org/abs/2109.05433
CPL
PublicOfficial implementation of our EMNLP 2022 paper "CPL: Counterfactual Prompt Learning for Vision and Language Models"ACLToolBox
Public