PCIe 4.0 Controller supporting Endpoint, Root Port, Switch, Bridge and advanced features
AI expands HBM footprint
By Gary Hilson, EETimes (January 20, 2022)
High bandwidth memory (HBM) is becoming more mainstream. With the latest iteration’s specifications approved, vendors in the ecosystem are gearing to make sure it can be implemented so customers can begin to design, test and deploy systems.
The massive growth and diversity in artificial intelligence (AI) means HBM is less than niche. It’s even become less expensive, but it’s still a premium memory and requires expertise to implement. As a memory interface for 3D-stacked DRAM, HBM achieves higher bandwidth while using less power in a form factor that’s significantly smaller than DDR4 or GDDR5 by stacking as many as eight DRAM dies with an optional base die which can include buffer circuitry and test logic.
Like all memory, HBM makes advances in performance improvement and power consumption with every iteration. A key change when moving to HBM3 from HBM2 will a 100% performance improvement in the data transfer rate from 3.2/3.6Gbps to 6.4Gbps max per pin, said Jinhyun Kim, principal engineer with Samsung Electronics’ memory product planning team.
A second fundamental change is a 50% increase in the maximum capacity from 16GB (8H) to 24GB (12H). Finally, HBM3 implements on-die error correction code as an industry-wide standard, which improves system reliability, Kim said. “This will be critical for the next generation of artificial intelligence and machine learning systems.”
![]() |
E-mail This Article | ![]() |
![]() |
Printer-Friendly Page |
Related News
- Flex Logix Expands Management Team To Meet Growing Demand For Its AI Inference and eFPGA Solutions
- Synopsys Expands Use of AI to Optimize Samsung's Latest Mobile Designs
- Mythic Expands Product Lineup with New Scalable, Power-Efficient Analog Matrix Processor for Edge AI Applications
- Siemens acquires Avatar, expands EDA footprint with innovative Place and Route technology
- Lattice Expands Ultra-Low Power sensAI Stack with Optimized Solutions for Always-On, On-Device AI