Stars
Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch
PyMuPDF is a high performance Python library for data extraction, analysis, conversion & manipulation of PDF (and other) documents.
MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.
A series of large language models developed by Baichuan Intelligent Technology
Hackable and optimized Transformers building blocks, supporting a composable construction.
ggml implementation of the baichuan13b model (adapted from llama.cpp)
Implementation of Nougat Neural Optical Understanding for Academic Documents
Fast and memory-efficient exact attention
CMMLU: Measuring massive multitask language understanding in Chinese
A 13B large language model developed by Baichuan Intelligent Technology
The official GitHub page for the survey paper "A Survey of Large Language Models".
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper, Ada and Blackwell GPUs, to provide better performance with lower memory…
Transformer related optimization, including BERT, GPT
Train transformer language models with reinforcement learning.
Ongoing research training transformer models at scale
Collections of vector search related libraries, service and research papers
Approximate nearest neighbor search with product quantization on GPU in pytorch and cuda
Automatically create Faiss knn indices with the most optimal similarity search parameters.
Library for 8-bit optimizers and quantization routines.
Fast and memory-efficient clustering
hnsw lib with hamming distance and uint32 coding
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.