-
Notifications
You must be signed in to change notification settings - Fork 78
Pull requests: intel/auto-round
Pull requests list
[fix]AttributeError: 'Autotuner' object has no attribute '_cache_lock'
#1414
opened Feb 5, 2026 by
xin3he
Loading...
3 of 9 tasks
refine qwen3_vl_moe experts forward
#1413
opened Feb 5, 2026 by
WeiweiZhang1
Loading...
2 of 9 tasks
better warning for not support mllm while quant_non_text_module
#1411
opened Feb 5, 2026 by
n1ck-guo
Loading...
1 of 9 tasks
update auto-round-kernel package name to auto-round-lib
#1410
opened Feb 5, 2026 by
chensuyue
Loading...
2 of 9 tasks
adapt vllm_ext to new extra config
#1409
opened Feb 5, 2026 by
mengniwang95
Loading...
1 of 6 tasks
0.10.0
Support Qwen3 Omni model quantization
#1404
opened Feb 4, 2026 by
lvliang-intel
•
Draft
2 of 9 tasks
support multiple device evaluation for activation quantized model
#1394
opened Feb 4, 2026 by
wenhuach21
Loading...
9 tasks
Optimize CPU RAM peak memory during quantization
#1386
opened Feb 3, 2026 by
lvliang-intel
Loading...
3 of 9 tasks
fix Qwen3-VL model auto_awq export, add auto_awq vllm ut
#1384
opened Feb 2, 2026 by
WeiweiZhang1
Loading...
3 of 9 tasks
Refactor module access to use PyTorch get/set_submodule API
#1365
opened Jan 29, 2026 by
scopophobic
Loading...
support hadamard transform for mxfp4 with rtn or autoround method.
#1349
opened Jan 27, 2026 by
lkk12014402
Loading...
refactor init of compressor
engineering
ready
only add when the PR is ready to merge
#1339
opened Jan 26, 2026 by
n1ck-guo
Loading...
1 of 9 tasks
Robust FP8 layer detection for ignore_layers (#1283)
#1289
opened Jan 15, 2026 by
scopophobic
Loading...
Fix ignore_layers not working for FP8 models
#1286
opened Jan 15, 2026 by
Copilot
AI
Loading...
11 tasks done
[WIP][refactor quanizers][step 1] refactor rtn and tuning
#1278
opened Jan 14, 2026 by
n1ck-guo
Loading...
add per-task lm_eval args for exprimental usage
#1017
opened Nov 11, 2025 by
WeiweiZhang1
Loading...
ProTip!
Filter pull requests by the default branch with base:main.
You can’t perform that action at this time.