Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

PQKD compresses CNN models via iterative pruning, performance recovery with knowledge distillation, and quantization-aware training, reducing model size by ~20× with minimal accuracy loss.

Notifications You must be signed in to change notification settings

rusuanjun007/PQKD

Open more actions menu

Repository files navigation

Pruning-Quantization with Knowledge Distillation(PQKD)

Introduction

PQKD is a method to compress a model by pruning and quantization with knowledge distillation. Through iterative pruning, performance recovering using knowledge distillation and followed by quantization-aware training (QAT), the PQKD successfully reduces the CNN-based model size by approximately 20 times while maintaining minimal degradation in accuracy. The channel adapters are inserted to match middle layer feature maps, solving the model heterogeneity problem caused by structured pruning.

PQKD

How to use

The PQKD is implemented in PyTorch. First pre-train the model in FP32 with fp32_pre_training.py, then run pruning_with_knowledge_distillation.py to iteratively pruning with knowledge distillation. Finally, run QAT_finetune.py to quantize the model.

Results

The PQKD achieves 20x compression with minimal accuracy degradation on [PEC datasets][https://www.kaggle.com/datasets/rusuanjun/pec-dataset]. The following table shows the results of ResNet50-1D and MobileNetV3 after pruning with knowledge distillation.

pkd

About

PQKD compresses CNN models via iterative pruning, performance recovery with knowledge distillation, and quantization-aware training, reducing model size by ~20× with minimal accuracy loss.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

Morty Proxy This is a proxified and sanitized view of the page, visit original site.