Carlos Azeredo Leme & Manuel Mota (Synopsys)
EDN (April 23, 2013)
Mobile multimedia devices process and combine audio signals from a variety of sources, including the baseband processor, Bluetooth enabled devices, and WiFi networks. The result is that today's smartphones and tablets are 'digital audio hubs' that must receive multiple asynchronous digital audio signals, synchronize them, and output them on loudspeakers or headsets.
Tablet/smartphone system-on-chips (SoCs) designers are faced with the challenge of implementing the complex audio mixing functions in the most cost-effective way possible. Traditionally, this function can be implemented in the application processor or in a dedicated audio processor; however, this is not the most cost-effective way to use the limited processing resources available to the system.
By integrating an audio analog codec that implements the 'audio hub' functionality and is able to process and mix audio signals from asynchronous sources, system designers can free the scarce main processor resources for more relevant tasks and simplify the system design, thus achieving a more effective solution.
This article will analyze:
- The benefits of having the audio codec in mobile multimedia systems operating as a digital audio hub to interconnect the different audio signal sources and destinations, each having independent clock domains
- How to synchronize and combine the various audio streams originated by different sources in the system, using built-in asynchronous sample rate converters (ASRCs).
By leveraging the latest improvements in audio codec IP, designers and system architects will be able to deliver the tightly integrated solutions that will make their SoCs stand out from the competition while minimizing costs.
Click here to read more ...