-
Beihang University
- Beijing, China
-
09:43
(UTC +08:00) - zhgyqc.site
- @zhgyqc_duguce
Highlights
Stars
Train transformer language models with reinforcement learning.
Ongoing research training transformer models at scale
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 500+ LLMs (Qwen3, Qwen3-MoE, Llama4, GLM4.5, InternLM3, DeepSeek-R1, ...) and 200+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, Llava, GLM4v, Ph…
Paper list for Efficient Reasoning.
dInfer: An Efficient Inference Framework for Diffusion Language Models
The official GitHub repository of the paper "Recent advances in large langauge model benchmarks against data contamination: From static to dynamic evaluation"
"what, how, where, and how well? a survey on test-time scaling in large language models" repository
[ACMMM 2025] Official Code of DetectAnyLLM: Towards Generalizable and Robust Detection of Machine-Generated Text Across Domains and Models
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
🚀🚀 「大模型」2小时完全从0训练26M的小参数GPT!🌏 Train a 26M-parameter GPT from scratch in just 2h!
A resource repository for machine unlearning in large language models
Awesome LLM pre-training resources, including data, frameworks, and methods.
A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.
Minimalistic large language model 3D-parallelism training
A Simple Framework of Small-scale LMMs for Video Understanding
A Framework of Small-scale Large Multimodal Models
MLNLP: Paper Picture Writing Code
小红书笔记 | 评论爬虫、抖音视频 | 评论爬虫、快手视频 | 评论爬虫、B 站视频 | 评论爬虫、微博帖子 | 评论爬虫、百度贴吧帖子 | 百度贴吧评论回复爬虫 | 知乎问答文章|评论爬虫
A powerful toolkit for compressing large models including LLM, VLM, and video generation models.
Existing Literature about Machine Unlearning
Official repository of Automated Privacy Information Annotation in Large Language Model Interactions