The report - Artificial Intelligence - Technology Adoption and Impact Report 2019 - suggests that within the AI segment, worldwide revenue from memory devices in AI applications will increase to $60 billion in 2025, up from $21 billion in 2019. The processor segment will expand slightly faster, growing to $68.5 billion in 2025, up from $22 billion in 2019.
This total tracks sales of semiconductor content in systems that run AI functions and these chips include memory and processing devices within systems that can run AI applications.
AI chips are used widely in various markets, including automotive, communication, computers, consumer electronics, industrial and healthcare. The largest single market for memory devices in AI applications is the computer segment, with sales rising to $66 billion in 2025, increasing at a 15.7 percent compound growth rate (CAGRs) from $27.5 billion in 2019. However, other segments will generate faster growth, including the communication, consumer electronics, industrial and healthcare sectors.
“Semiconductors represent the foundation of the AI supply chain, providing the essential processing and memory capabilities required for every artificial intelligence application on earth,” said Luca De Ambroggi, senior research director for AI at IHS Markit | Technology. “AI is already propelling massive demand growth for microchips. However, the technology also is changing the shape of the chip market, redefining traditional processor architectures and memory interfaces to suit new performance demands.”
The report says that a growing number of start-ups are aiming to offer completely new architectures that will challenge the market supremacy of traditional devices used for AI processing, such as graphics processing units (GPUs), field programmable gate arrays (FPGAs), microprocessors (MPUs), microcontrollers (MCUs), and digital signal processors (DSPs). These new architectures include capabilities such as integrated vector-processing, which can accelerate deep-learning tasks.
Moreover, the introduction of AI-related capabilities into various devices means that these traditional classes of processors are evolving to the point where they are no longer recognisable as distinct categories.
“The old definitions of what makes an MPU, DSP or MCU are beginning to blur in the AI era, as each type of device adds cores with different core functions,” De Ambroggi said. “Increasingly, designers of AI-enabled systems are using highly integrated heterogenous processing solutions, such application-specific integrated circuits (ASICs) and system-on-chip (SoC) solutions.
"With processor makers offering turnkey, heterogenous processing solutions using these ASICs and SOCs, it makes less difference to system designers whether their AI algorithm is executed on a GPU, CPU or DSP.”
Advanced AI technologies, specifically deep-learning algorithms, require huge amounts of high-bandwidth volatile memory to perform properly but this has been driving power consumption to unsustainable levels.
In response the industry has been looking at a number of alternative approaches including: a new processor architecture wherein the memory is closer to the computational core, reducing the burden of data movement and enabling high processing parallelism with dedicated memory cells for each processing core; moving the early stages of data computation into the memory, a technique called processing into memory (PIM). PIM delivers similar benefits as those mentioned above and, finally, identifying new memory technologies that can enable new approaches with easy back-end silicon integration, volatile performance, a non-volatile capability, low pico-joule per byte or a new and a fast input/output (I/O) interface.
To download the report follow the link below.