RaiderChip Hardware NPU adds Falcon-3 LLM to its supported AI models
The model, launched by Abu Dhabi’s Technology Innovation Institute (TII), runs seamlessly in RaiderChip’s AI accelerator and its FPGA based demonstrator.
Spain, December 19th, 2024 -- On December 17th, the Technology Innovation Institute (TII) of Abu Dhabi (United Arab Emirates) launched Falcon 3, a new open-source generative AI model available in four sizes: 1B, 3B, 7B, and 10B parameters. Categorized as a small model by its size, Falcon 3 delivers outstanding performance when compared to competing LLMs in its category.
Thanks to RaiderChip technology’s ease of use, implementing support for new LLM AI models is straight-forward and transparent. Only a firmware update is needed to accelerate Falcon-3 inference fully inside its hardware NPU.
RaiderChip NPU running, fully standalone and offline, the Falcon-3 1B LLM with BF16 weights
Falcon 3’s design, optimized for lightweight infrastructure efficiency, is particularly appealing to RaiderChip as it aligns with the company’s strategy to offer generative AI solutions that are embeddable on low-cost hardware with limited resources. According to Victor Lopez, RaiderChip’s CTO, “The fact that Falcon models are available in various sizes, including quantized versions, allows us to provide our customers with fast and efficient inference across a wide range of edge and embedded devices using our state-of-the-art NPU, which provides Generative AI in fully offline and standalone devices.”
The introduction of small models, capable of delivering extraordinary reasoning, comprehension, and generation capabilities across multiple languages, enhances the scope and quality of generative AI applications based on FPGAs. These include conversational assistants, recommendation systems, and data analysis platforms, among others —an expanding list expected to grow significantly in the coming months and years.
RaiderChip’s commitment to low-cost FPGAs contributes to the diversification and democratization of access to advanced artificial intelligence. This aligns with Falcon 3’s philosophy of offering powerful and accessible models to all users. Thanks to RaiderChip’s NPU, Falcon 3 can now run locally, offline, and on-premises across a greater number of new devices, unlocking new opportunities for generative AI at the edge.
Companies interested in trying the Generative AI NPU may reach out to RaiderChip for access to our FPGA demo or a consultation on how our IP cores can accelerate their AI workloads.
More information at https://raiderchip.ai/technology/hardware-ai-accelerators
|
Related News
- RaiderChip launches its Generative AI hardware accelerator for LLM models on low-cost FPGAs
- RaiderChip NPU for LLM at the Edge supports DeepSeek-R1 reasoning models
- RaiderChip unveils its fully Hardware-Based Generative AI Accelerator: The GenAI NPU
- Ceva and Edge Impulse Join Forces to Enable Faster, Easier Development of Edge AI Applications
- AMD Completes Acquisition of Silo AI to Accelerate Development and Deployment of AI Models on AMD Hardware
Breaking News
- GUC Announces Tape-Out of the World's First HBM4 IP on TSMC N3P
- VeriSilicon Launches ISP9000: The Next-Generation AI-Embedded ISP for Intelligent Vision Applications
- Frontgrade Gaisler Launches New GRAIN Line and Wins SNSA Contract to Commercialize First Energy-Efficient Neuromorphic AI for Space Applications
- Latest intoPIX JPEG XS Codec Powers FOR-A's FA-1616 for Efficient IP Production at NAB 2025
- BrainChip Partners with RTX's Raytheon for AFRL Radar Contract
Most Popular
- Movellus and RTX's SEAKR Engineering Collaborate on Advancing Mission-Critical ASICs
- Intel brings 3nm production to Europe in 2025
- Silicon-Proven MIPI CSI-2 & DSI-2 Tx/Rx IP Cores for your Camera & Display SoCs
- Eliyan Sets New Standard for Chiplet Interconnect Performance with Latest PHY Delivering Data Rate of 64Gbps on 3nm Process Using Standard Packaging
- EnSilica Opens Second Brazilian Design Centre Following Multimillion Pound Design and Manufacturing Contract Win
![]() |
E-mail This Article | ![]() |
![]() |
Printer-Friendly Page |