Liveness Analysis Helps Save Gigabytes of Memory Usage for AI Inference

Figure 1: Memory footprint normalized to total tensor size. The smaller, the better.
Figure 2: IR of LeNet, transformed from ONNX by ONNC.
Figure 3: Memory allocation result for LeNet.

--

--

--

The Open Neural Network Compiler (ONNC), a compiler that connects Open Neural Network Exchange Format (ONNX) to every deep learning accelerator (DLA).

Love podcasts or audiobooks? Learn on the go with our new app.

Recommended from Medium

From Jupyter Notebooks to Real-life: MLOps

Implementing Facial Recognition via Transfer Learning

Extracting image features without Deep Learning

Image Captioning with CLIP and GPT

Build and deploy Image Classification web app using Django on Heroku (Part-1)

AZURE MACHINE LEARNING

Deep Double Descent: when more data and bigger models are a bad thing

Decision Trees

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
ONNC

ONNC

The Open Neural Network Compiler (ONNC), a compiler that connects Open Neural Network Exchange Format (ONNX) to every deep learning accelerator (DLA).

More from Medium

Bias in Machine Learning

Choosing the number of clusters in K-means learning algorithm

Are they similar?

Tuberculosis Detection with A.I

png