Simulate, test, and deploy machine learning and deep learning models to edge devices and embedded systems. Generate code for complete AI applications, including pre-processing and post-processing algorithms.
With MATLAB® and Simulink®, you can:
- Generate optimized C/C++ and CUDA code for deployment to CPUs and GPUs
- Generate synthesizable Verilog and VHDL code for deployment to FPGAs and SoCs
- Accelerate inference with hardware-optimized deep learning libraries, including oneDNN, Arm Compute Library, and TensorRT
- Incorporate pre-trained TensorFlow Lite (TFLite) models into applications deployed to hardware
- Compress AI models for inference on resource-constrained hardware with tools for hyperparameter tuning, quantization, and network pruning
CPUs and Microcontrollers
Generate portable, optimized C/C++ code from trained machine learning and deep learning models with MATLAB Coder™ and Simulink Coder™. Optionally include calls to vendor-specific libraries for deep learning inference in the generated code, such as oneDNN and Arm® Compute Library.
GPUs
Generate optimized CUDA® code for trained deep learning networks with GPU Coder™. Include pre-processing and post-processing along with your networks to deploy complete algorithms to desktops, servers, and embedded GPUs. Use NVIDIA® CUDA libraries, such as TensorRT™ and cuDNN, to maximize performance.
FPGAs and SoCs
Prototype and implement deep learning networks on FPGAs and SoCs with Deep Learning HDL Toolbox™. Program deep learning processors and data movement IP cores with pre-built bitstreams for popular FPGA development kits. Generate custom deep learning processor IP cores and bitstreams with HDL Coder™.
AI Model Compression
Reduce memory requirements for machine learning and deep learning models with size-aware hyperparameter tuning and quantization of weights, biases, and activations. Minimize the size of a deep neural network by pruning insignificant layer connections.