Stars
🤖 AgentVerse 🪐 is designed to facilitate the deployment of multiple LLM-based agents in various applications, which primarily provides two frameworks: task-solving and simulation
An Open-source Framework for Data-centric, Self-evolving Autonomous Language Agents
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Build resilient language agents as graphs.
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
xDiT: A Scalable Inference Engine for Diffusion Transformers (DiTs) with Massive Parallelism
A high-throughput and memory-efficient inference and serving engine for LLMs
This is a collection of our NAS and Vision Transformer work.
LAVIS - A One-stop Library for Language-Vision Intelligence
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
DeepSeek-VL: Towards Real-World Vision-Language Understanding
A Framework of Small-scale Large Multimodal Models
Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models (CVPR 2024 Highlight)
【AAAI2025】MambaPro: Multi-Modal Object Re-Identification with Mamba Aggregation and Synergistic Prompt
CVPR 2025 | Every SAM Drop Counts: Embracing Semantic Priors for Multi-Modality Image Fusion and Beyond
[CVPR 2025] SegMAN: Omni-scale Context Modeling with State Space Models and Local Attention for Semantic Segmentation
[AAAI 2025] Official PyTorch implementation of "TinySAM: Pushing the Envelope for Efficient Segment Anything Model"
[ICML 2024] Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
[NeurIPS2025 Spotlight 🔥 ] Official implementation of 🛸 "UFO: A Unified Approach to Fine-grained Visual Perception via Open-ended Language Interface"
[CVPR 2025 Highlight] Official code for paper "Mamba as a Bridge: Where Vision Foundation Models Meet Vision Language Models for Domain-Generalized Semantic Segmentation"
📄 Easily create your resume with Markdown on VSCode / Typora / Obsidian
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
每个人都能看懂的大模型知识分享,LLMs春/秋招大模型面试前必看,让你和面试官侃侃而谈
2025年最新科学上网,vpn机场推荐,支持shadowrocket,ss, ssr, v2ray, trojan, clash,clashr,需要自取(每日更新)