Many models that achieve state-of-the-art performance on benchmarks and when practically applied are computationally expensive and memory intensive. However, sparsification methods can be hard to implement. To make it easier to compress big deep learning models, Alibaba has released TinyNeuralNetwork. TinyNeuralNetwork can enable deployment on-device by deploying model compression techniques like pruning, quantization, and model conversion.