模型压缩指南
概述
支持量化和剪枝两种压缩方法。
快速开始
量化
python
from med_core.utils.compression import quantize_model
# 动态量化
quantized_model = quantize_model(model, method="dynamic")剪枝
python
from med_core.utils.compression import prune_model
# 非结构化剪枝
pruned_model = prune_model(model, amount=0.3)完整压缩
python
from med_core.utils.compression import compress_model
compressed_model = compress_model(model, quantize=True, prune=True)参考
med_core/utils/compression.py- 本页代码片段可直接作为起点;当前仓库未单独提供
examples/model_compression_demo.py