HC2021.C1.4 Intel Arijit
HC2021.C1.4 Intel Arijit
HC2021.C1.4 Intel Arijit
Arijit Biswas
Intel Senior Principal Engineer
New Standard for
Data Center Architecture
Cache &
Scalar Data Parallel Memory Sub- Intra/Inter
Performance Performance System Arch Socket Scaling
PCIe 5.0
Next Gen Quality of
Service Capabilities
Broad WL/Usage
Support and
Low Jitter Optimizations
Fast VM Migration
Architecture
Elasticity &
Efficient Data Infrastructure &
Consolidation Performance Center Framework
& Orchestration Consistency Utilization Overhead
Compute IP Cores
Acceleration
Seamless Integration of Engines
PCIe Gen
I/O IP CXL1.1
5
UPI 2.0
Architecture
Improvements for DC Half- Precision Float New Instructions
Workloads & Usages HFNI Support for FP16 - higher throughput lower precision
Cache CLDEMOTE
Management Proactive placement of cache contents
Sapphire Rapids
Utilization Without
Acceleration
Acceleration Engines
Increasing effectiveness of cores,
by enabling offload of common mode tasks via
seamlessly integrated acceleration engines Core Core Core Core
With Acceleration
Native Dispatch, Signaling & Synchronization from User Space
Accelerator interfacing Architecture
Utilization
Coherent, Shared Memory Space
Between Cores & Acceleration Engines
Core Core Core Core Accel. Accel.
Concurrently shareable
Processes, containers and VMs
Critical Workloads Common Mode Tasks Additional Workload Capacity
Intel® Data Streaming Acceleration Engine
Optimizing streaming data movement and transformation operations
14%
up to
Results have been estimated or simulated based on testing on pre-production hardware and software.
For workloads and configurations visit www.intel.com/ArchDay21claims . Results may vary
Intel® Quick Assist Technology Acceleration Engine
Accelerating Cryptography and Data De/Compression
100%
up to
160Gb/s Compression +
98%
160Gb/s De-compression additional
workload capacity
Fused Operations
after QAT offload
Improved RAS
2 Modes
HBM
HBM DDR5
HBM DDR5
HBM Flat Mode enables flat memory HBM Caching Mode allows HBM
regions with HBM & DRAM to act as DRAM backed cache
Sapphire Rapids - Architected for AI
AI has become ubiquitous across usages – AI performance required in all tiers of computing
Goal 2048
Enable efficient usage of AI across all services deployed on
elastic general-purpose tier by delivering many times more AI
Available and integrated with AVX-512 (2xFMA) FP32 AVX-512 (2xFMA) INT8
AMX (TMUL) BF16 AMX (TMUL) INT8
industry-relevant frameworks & libraries
Results have been simulated. For workloads and configurations
visit www.intel.com/ArchDay21claims . Results may vary
Sapphire Rapids - Built for elastic computing models - microservices
>80% of new cloud-native and SaaS applications are expected to be built as microservices
Microservices Performance
Goal
+69%
Reduced
Infrastructure
Overhead
General Purpose
Sapphire Rapids
Multi Tile SoC for Physically Tiled, & Dedicated Biggest Leap in Data Center Capabilities
Scalability Logically Monolithic Acceleration Engines in over a Decade
Enhanced
DDR 5 & Virtualization
HBM PCIe 5.0 Capabilities
22