-
Cornell University
- NYC
- https://giladturok.github.io
Stars
The Family of Diffusion Protein Language Models (DPLM)
VeOmni: Scaling Any Modality Model Training with Model-Centric Distributed Recipe Zoo
[COLM '25] Single-Pass Document Scanning for Question Answering
CLiFF (Clustering & Language model integration for FAQ Formation)
A repository of links with advice related to grad school applications, research, phd etc
Thermodynamic Hypergraphical Model Library in JAX
Don't just regulate gradients like in Muon, regulate the weights too
[NeurIPS 2025] Encoder-Decoder Diffusion Language Models for Efficient Training and Inference
Accessible large language models via k-bit quantization for PyTorch.
LLM model quantization (compression) toolkit with hw acceleration support for Nvidia CUDA, AMD ROCm, Intel XPU and Intel/AMD/Apple CPU via HF, vLLM, and SGLang.
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
TL;DR: We only have one life. Let's stop wasting it on YouTube shorts.
Fast and memory-efficient exact attention
Minimal and annotated implementations of key ideas from modern deep learning research.
Mirage Persistent Kernel: Compiling LLMs into a MegaKernel
A comprehensive JAX/NNX library for diffusion and flow matching generative algorithms, featuring DiT (Diffusion Transformer) and its variants as the primary backbone with support for ImageNet train…
Serialize JAX, Flax, Haiku, or Objax model params with 🤗`safetensors`
📊 Save matplotlib figures as TikZ/PGFplots for smooth integration into LaTeX.
Official PyTorch implementation for ICLR2025 paper "Scaling up Masked Diffusion Models on Text"
Awesome Reasoning LLM Tutorial/Survey/Guide
Post-training with Tinker