Model Compression¶
| Quantization | Reducing precision from Float64 to Int8 |
| Pruning | Removing unnecessary aspects of the model Removing neurons in ANN |
2024-01-24
| Quantization | Reducing precision from Float64 to Int8 |
| Pruning | Removing unnecessary aspects of the model Removing neurons in ANN |