Comments (4)
Hello sijeh,
Sorry for the late reply and thanks for your advice, I am planning to revise the code and provide the resnet-18 checkpoints. I will comment to you when the code is updated and the checkpoints are uploaded.
from apot_quantization.
Thx.
from apot_quantization.
Hi sijeh,
I just uploaded the ckpts for the 4bit ResNet-18 and the new codes for the APoT Quantization! Here are the changes:
- quantization function now does not manually overwrite the gradients
- You can specify the bitwidth when initializing the model
- Checkpoints for 4-bit, 3-bit ResNet-18 are uploaded. More ckpts will come in a few days
- Tensorboard log file is also provided in the events dir
- New Hyperparams: Please note that we have changed the hyper-params configuration. For ResNet-18-4bit, LR is set to 0.01 and scaled by 0.1 for all parameters including clipping thresholds. Weight decay is set to 1e-4 for all parameters. Please also note that 4-bit model are initialized by 5-bit quantized model. (If you do not pre-train a 5-bit model, it's fine to directly initialize it from full precision model. However, we recommend you to progressively initialize the low-bit model, e.g. 2-bit).
Regarding your question about batch size:
Theoretically speaking, LR is proportional to the batch size because lower batch size causes more training iterations. Therefore, you may use 0.01*192/1024 as your base LR.
If you still have further question, please do not hesitate to comment here.
from apot_quantization.
Hi yhhhli,
Thanks for your detailed reply and updates of the opensource code and pretrained model, all the things going to be right since I re-downloaded and unzip ImageNet dataset.
from apot_quantization.
Related Issues (20)
- Why the size of Res20_2bit is the same as Res20_32bit? HOT 1
- uniform_quantization HOT 1
- Size and accuracy HOT 5
- Lightning Integration
- Technical details HOT 2
- about uniform quantization HOT 2
- about CIFAR10 part main.py resume function HOT 2
- the precision a4w4 of training MobilenetV2 is nearly 0 HOT 4
- a4w4 Resnet18 is 1.7% lower than that in the paper?
- The MUL unit of APOT HOT 1
- Need Suggestion
- Some results about resnet20 on cifar10
- quantization bit of apot HOT 2
- about training time
- difference between paper and code in quan_layer HOT 2
- calculate MAC
- Hyper-Params on MobileNet_V2 HOT 1
- The migration of this QAT function? HOT 5
- NaN loss for 8bit HOT 1
- Differences between quant_layer.py HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from apot_quantization.