Model Compression¶
| Quantization | Reducing precision from Float64 to Int8 |
| Pruning | Removing unnecessary aspects of the model Removing neurons in ANN |
| Quantization | Reducing precision from Float64 to Int8 |
| Pruning | Removing unnecessary aspects of the model Removing neurons in ANN |