Optimizing AI and Machine Learning with eFPGAs
By Cheng Wang, Flex Logix, Inc.
August 6th, 2018, eecatalog.com
Why the performance and flexibility offered by eFPGA is turning out to be a game changer for anyone designing AI and machine learning and struggling to meet the compute demands.
The market for artificial intelligence (AI) and machine learning applications has been growing substantially over the last several years. Designers have a tough row to hoe when it comes to satisfying these applications’ seemingly insatiable compute hunger. They are finding that traditional Von Neumann processor architectures are not optimal solutions for the neural networks fundamental to AI and machine learning.
When GPUs are used to train neural networks, they require floating pointing math that is very compute intensive. However, using integer math for inference, designers can speed computation by turning to FPGAs for neural network processing. Many companies are starting to recognize this, with Microsoft’s Project Brainwave, which uses FPGA chips to accelerate AI, as a perfect example.
E-mail This Article | Printer-Friendly Page |
|
Flex Logix Technologies, Inc. Hot IP
Related Articles
- Performance Evaluation of machine learning algorithms for cyber threat analysis SDN dataset
- Optimizing Electronics Design With AI Co-Pilots
- Exploring Machine Learning testing and its tools and frameworks
- Artificial Intelligence (AI) utilizing deep learning techniques to enhance ADAS
- An overview of Machine Learning pipeline and its importance
New Articles
- The Rise of RISC-V and ISO 26262 Compliance
- Synopsys 3DIO Solution for Multi-Die Integration (2.5D/3D)
- SoC NoCs: Homegrown or Commercial Off-the-Shelf?
- From a Lossless (~1.5:1) Compression Algorithm for Llama2 7B Weights to Variable Precision, Variable Range, Compressed Numeric Data Types for CNNs and LLMs