Deep Learning HDL Toolbox

 

Deep Learning HDL Toolbox

Prototype and deploy deep learning networks on FPGAs and SoCs

Screenshots of prototyping and implementing deep learning networks on FPGAs for edge deployment.
Block diagram of programmable deep learning processor.

Using an FPGA Based Deep Learning Processor

The toolbox includes a deep learning processor featuring generic deep learning layers controlled by scheduling logic. This processor performs  FPGA-based inferencing of networks using Deep Learning Toolbox.

Detection of a white truck with a bounding box.

Compiling and Deploying Networks

Compile a deep learning network into a set of instructions to be run by the deep learning processor. Deploy your network to the FPGA and run prediction while capturing actual on-device performance metrics.

Getting Started with Prebuilt Bitstreams

Rapidly prototype a long short-term memory (LSTM) network using available pre-built bitstreams. Customize bitstream configuration to meet resource use requirements.

Code snippet with Xilinx bitstream, connected to an FPGA development board.

Deploying Networks to FPGAs

Use Deep Learning Toolbox to develop RNN and CNN networks or import a network. Then program an FPGA using the deploy command, deploying to AMD or Intel FPGAs.

While loop MATLAB code for predict call.

Running FPGA-Based Inferencing in MATLAB Applications

Run a complete application in MATLAB, including your test bench, preprocessing and post-processing algorithms, and the FPGA-based deep learning inferencing. A single MATLAB command, predict, performs the inferencing on the FPGA and returns results to the workspace in MATLAB.

Profiling FPGA Inferencing and Tuning Network Designs

Using the profile metrics, tune your network configuration by measuring layer-level latency as you run predictions on the FPGA to find performance bottlenecks.

Compressing Deep Learning Network for FPGA Deployment

Reduce resource utilization by compressing your deep learning network with quantization and pruning. Analyze tradeoffs between accuracy and resource utilization using the Model Quantization Library support package.

Customizing the Deep Learning Processor Configuration

Specify hardware architecture options for implementing the deep learning processor, such as the number of parallel threads or maximum layer size.

Screenshot of custom bitstream coding.

Generating Synthesizable RTL and IP Cores

Use HDL Coder to generate synthesizable RTL from the deep learning processor. Generate IP core with standard AXI interfaces for integration into AMD and Intel SoC designs.