-
Notifications
You must be signed in to change notification settings - Fork 72
Pull requests: intel/auto-round
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
add blockwise way in pure rtn to ease module replacements
#1320
opened Jan 22, 2026 by
wenhuach21
Loading…
18 tasks
[GGUF] using quant_nontext_module to control whether quant vision model
#1317
opened Jan 22, 2026 by
n1ck-guo
Loading…
8 of 18 tasks
Delay materializing the replaced model weights until quantization
#1307
opened Jan 21, 2026 by
yiliu30
Loading…
5 of 6 tasks
Optimize FP8 layer conversion by skipping weight initialization
#1295
opened Jan 16, 2026 by
Copilot
AI
Loading…
Robust FP8 layer detection for ignore_layers (#1283)
#1289
opened Jan 15, 2026 by
scopophobic
Loading…
Fix ignore_layers not working for FP8 models
#1286
opened Jan 15, 2026 by
Copilot
AI
Loading…
11 tasks done
[WIP][refactor quanizers][step 1] refactor rtn and tuning
#1278
opened Jan 14, 2026 by
n1ck-guo
Loading…
ProTip!
Mix and match filters to narrow down what you’re looking for.