Stars
An extremely fast Python linter and code formatter, written in Rust.
A Unified and Flexible Inference Engine with Hybrid Cache Acceleration and Parallelism for 🤗Diffusers.
An extremely fast Python package and project manager, written in Rust.
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Entropy Based Sampling and Parallel CoT Decoding
Curating resources for learning how to trade
Prov-GigaPath: A whole-slide foundation model for digital pathology from real-world data
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
A dummy's guide to setting up (and using) HPC clusters on Ubuntu 22.04LTS using Slurm and Munge. Created by the Quant Club @ UIowa.
Prometheus exporter for performance metrics from Slurm.
Singularity image for a deep learning (pytorch) environment + GPU support
Port of OpenAI's Whisper model in C/C++
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
Official inference library for Mistral models
Pathology Language and Image Pre-Training (PLIP) is the first vision and language foundation model for Pathology AI (Nature Medicine). PLIP is a large-scale pre-trained model that can be used to ex…
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
stjordanis / MetaGPT
Forked from FoundationAgents/MetaGPT🌟 The Multi-Agent Framework: Given one line Requirement, return PRD, Design, Tasks, Repo
CLI platform to experiment with codegen. Precursor to: https://lovable.dev
You like pytorch? You like micrograd? You love tinygrad! ❤️
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch
StableLM: Stability AI Language Models