-
Meta AI
- Menlo Park, CA
- https://mktal.github.io/
- in/xiaochengt
- kai.xtang
Stars
Adds "modifier key + mouse drag" move and resize to OSX
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
AI & parametric QR code generator. AI & 参数化二维码生成器。https://qrbtf.com
800,000 step-level correctness labels on LLM solutions to MATH problems
The official implementation of “Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training”
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
A guidance language for controlling large language models.
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Streamlining reinforcement learning with RLOps. State-of-the-art RL algorithms and tools, with 10x faster training through evolutionary hyperparameter optimization.
High-quality single file implementation of Deep Reinforcement Learning algorithms with research-friendly features (PPO, DQN, C51, DDPG, TD3, SAC, PPG)
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
4 bits quantization of LLaMA using GPTQ
Lightweight wrapper of the official ChatGPT API in your terminal
Task-based datasets, preprocessing, and evaluation for sequence models.
Train to 94% on CIFAR-10 in <6.3 seconds on a single A100. Or ~95.79% in ~110 seconds (or less!)
An open source implementation of CLIP.
(NeurIPS '21 Spotlight) IQ-Learn: Inverse Q-Learning for Imitation
🦜🔗 The platform for reliable agents.
Train transformer language models with reinforcement learning.
A modular RL library to fine-tune language models to human preferences
ASDL: Automatic Second-order Differentiation Library for PyTorch