This repository has been archived by the owner on Sep 18, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 1.8k
[20220524] Compression Roadmap 2022
J-shang edited this page May 26, 2022
·
3 revisions
- The method of simulating the compression effect is to replace some nodes in the graph with the wrapped ones. But note that sometimes the method of wrapping only the current node is not equivalent to the actual compressed effect.
- base
- evaluator (handle train, validate, hook, patch...)
- api design
- config list refactor
- specify compression target (input, output, weight, ...)
- specify compression algo (include related parameters, such as sparse pattern, quant bit)
- Support for variable compression targets
- compressor & wrapper refactor, provide a unified interface for parsing config list.
- basic pruner refactor & quantizer design
- Super compressor? most existed basic pruner/quantizer can implement by config super compressor?
- universal wrapper
- evaluator (handle train, validate, hook, patch...)
- pruning
- refactor sparse pattern
- metric calculator & sparsity allocator
- migrate to evaluator
- refactor sparse pattern
- quantization
- refactor design (key consideration: experiment, evaluator, wrapper, conv-bn-fusion)
- experiment
- wrap tuner as strategy
- search space generator
- support more pruners & quantizers
- a good strategy (how to search in search space)
- wrap tuner as strategy
- speedup
- mask propagation stands alone as a module
- quantization speedup supports more backend
- benchmark
- visualization
This wiki is a journal that tracks the development of NNI. It's not guaranteed to be up-to-date. Read NNI documentation for latest information: https://nni.readthedocs.io/en/latest/