IDC's AI and Generative AI Infrastructure Stacks and Deployments service provides qualitative and quantitative insights on the infrastructure and infrastructure-as-a-service stacks for predictive AI and generative AI (GenAI) workloads. IDC defines an infrastructure stack as an integrated set of hardware and software platforms, systems, and technologies optimized for specific outcomes. IDC defines deployments to include shared and dedicated tenancy, cloud and noncloud deployments, capex and opex procurements, as well as on-premises, collocated, hosted, and cloud services. The service offers analyst perspectives on which infrastructure stacks, deployments, and consumption models are best suited for the myriad of AI and GenAI use cases. Specific focus includes infrastructure needs for AI and GenAI data preparation, AI model training, retraining, fine-tuning, optimization, and AI inferencing.
AI and Generative AI Infrastructure Stacks and Deployments
Related Links
Meet the Experts
Photo of Ashish Nadkarni
Ashish Nadkarni
Group Vice President and General Manager, Worldwide Infrastructure Research
Photo of Heather West, PhD
Heather West, PhD
Research Manager, Performance Intensive Computing, Worldwide Infrastructure Research
Photo of Peter Rutten
Peter Rutten
Research Vice-President, Performance Intensive Computing, Worldwide Infrastructure Research
Markets and Subjects Analyzed
- Commercial and self-built infrastructure stacks for AI and GenAI workloads and use cases
- Compute (processor and accelerator architectures, computing platforms and systems, operating environments, virtualization, and containerization software)
- Storage systems, data persistence mechanisms, organization, access, and connectivity
- Software that enables optimized access to the hardware) as well as the NVIDIA CUDA, OpenCL, and so forth
- Cloud and noncloud approaches for AI and GenAI use cases
Core Research
- Market Taxonomy for Infrastructure and Infrastructure-as-a-Service Stacks for AI and GenAI
- Market Size and Forecast for AI and GenAI Infrastructure and Infrastructure as a Service
- End-User Adoption Trends, Use Cases, and Evolving Application Requirements
- Processor and Coprocessor/Accelerator Trends for AI and GenAI Use Cases
- Commercial and Open Source File, Object, and Block Scale-Up/Scale-Out Platforms and Systems
In addition to the insight provided in this service, IDC may conduct research on specific topics or emerging market segments via research offerings that require additional IDC funding and client investment.
Key Questions Answered
- How big is the infrastructure and infrastructure-as-a-service market for AI and GenAI workloads?
- What are the infrastructure hardware and software requirements imposed by AI and GenAI workloads?
- What are some of the data life-cycle challenges associated with AI and GenAI workloads?
- What are the optimal compute and storage configurations for AI and GenAI workloads?
- What is the role of accelerated computing (GPUs, FPGAs, ASICs, manycore processors, and emerging acceleration technologies), NVMe, tiering, de-duplication, and compression as they are related to AI and GenAI workloads?
Companies Covered
- Advanced Micro Devices, Inc.
- Alibaba Group Holding Limited
- Amazon Web Services Inc.
- Baidu, Inc.
- Broadcom Inc.
- Cisco Systems Inc.
- DataDirect Networks, Inc.
- Dell Technologies Inc.
- Google LLC
- Hewlett Packard Enterprise
- Hortonworks, Inc.
- Huawei Technologies Co., Ltd.
- IBM
- Inspur Group Co., Ltd.
- Intel Corporation
- Juniper Networks, Inc.
- Lenovo Group Limited
- Meta Platforms Inc.
- Microsoft Corporation
- NVIDIA Corporation
- NetApp, Inc.
- Oracle Corporation
- Penguin Computing Inc.
- Pure Storage, Inc.
- SAP SE
- SAS Institute Inc.
- SambaNova Systems, Inc.
- Super Micro Computer Inc.
- Symantec Corporation
- Tencent Holdings Limited