GPU Architecture and NVIDIA & AMD Technologies Overview¶
CPU vs. GPU: Architectural Differences¶
1. What is a primary advantage of GPUs over CPUs for high-performance computing?
- A. Higher clock speeds
- B. Optimized for sequential tasks
- C. Large number of cores optimized for parallel processing
- D. Smaller physical size
Click to reveal the answer
Answer: C. Large number of cores optimized for parallel processing2. True or False: CPUs generally have thousands of cores optimized for handling many tasks in parallel.
Click to reveal the answer
Answer: FalseGPU Core Organization: GPCs, SMs, and TPCs¶
3. In NVIDIA GPUs, what does SM stand for?
- A. Shared Memory
- B. Single Matrix
- C. Streaming Multiprocessor
- D. Spatial Memory
Click to reveal the answer
Answer: C. Streaming Multiprocessor4. True or False: NVIDIA GPUs operate in a Single Instruction, Multiple Threads (SIMT) fashion, allowing each SM to manage thousands of threads simultaneously.
Click to reveal the answer
Answer: True5. What is the primary purpose of a Texture Processor Cluster (TPC) in NVIDIA GPUs?
- A. To manage global memory
- B. To process texture data
- C. To handle thread scheduling
- D. To control data flow between CPUs and GPUs
Click to reveal the answer
Answer: B. To process texture dataGPU Memory Hierarchy¶
6. Which type of memory in NVIDIA GPUs is cached on-chip and optimized for frequently accessed constants?
- A. Global Memory
- B. Local Memory
- C. Constant Memory
- D. Texture Memory
Click to reveal the answer
Answer: C. Constant Memory7. True or False: Shared memory in NVIDIA GPUs is on-chip memory within each SM, allowing for faster access times and efficient data reuse.
Click to reveal the answer
Answer: True8. In GPU memory hierarchy, what type of memory is primarily used for handling large datasets and is accessible by both host and device?
- A. Local Memory
- B. Global Memory
- C. Texture Memory
- D. Register Memory
Click to reveal the answer
Answer: B. Global MemorySM Architecture and Execution Model¶
9. How many threads are grouped together in a warp on NVIDIA GPUs?
- A. 8
- B. 16
- C. 32
- D. 64
Click to reveal the answer
Answer: C. 3210. True or False: Cooperative Thread Arrays (CTAs) organize threads into blocks that execute in parallel on the GPU.
Click to reveal the answer
Answer: TrueNVIDIA Microarchitectures¶
11. Which NVIDIA microarchitecture introduced tensor cores specifically designed for AI workloads?
- A. Maxwell
- B. Volta
- C. Kepler
- D. Fermi
Click to reveal the answer
Answer: B. Volta12. True or False: Compute capability defines the feature set available for each NVIDIA GPU architecture.
Click to reveal the answer
Answer: TrueUnified Memory¶
13. What benefit does NVIDIA's Unified Memory provide?
- A. Increases GPU clock speeds
- B. Reduces memory transfer bottlenecks between CPU and GPU
- C. Optimizes GPU temperature control
- D. Improves network communication
Click to reveal the answer
Answer: B. Reduces memory transfer bottlenecks between CPU and GPU14. True or False: Unified Memory is particularly beneficial for applications where data must be shared between the CPU and GPU.
Click to reveal the answer
Answer: TrueNVLink and NVSwitch¶
15. NVLink and NVSwitch are designed to improve data transfer speed between:
- A. CPU and RAM
- B. GPU and Storage
- C. GPU and Network
- D. GPU and GPU
Click to reveal the answer
Answer: D. GPU and GPU16. True or False: NVLink outperforms traditional PCIe interconnects in multi-GPU configurations, improving data throughput.
Click to reveal the answer
Answer: TrueAMD GPU Architecture and Technologies Overview¶
17. In AMD GPUs, cores are organized into groups called:
- A. Compute Units (CUs)
- B. Stream Multiprocessors (SMs)
- C. Processing Arrays (PAs)
- D. Warp Clusters
Click to reveal the answer
Answer: A. Compute Units (CUs)18. True or False: AMD's wavefronts contain 64 threads, comparable to NVIDIA's warps which contain 32 threads.
Click to reveal the answer
Answer: TrueAMD ROCm Platform¶
19. What is HIP in the context of AMD's ROCm platform?
- A. A graphics library for rendering
- B. A parallel processing model exclusive to CPUs
- C. A C++ runtime API allowing code portability between AMD and NVIDIA GPUs
- D. A type of memory in AMD GPUs
Click to reveal the answer
Answer: C. A C++ runtime API allowing code portability between AMD and NVIDIA GPUs20. True or False: MIOpen is an AMD library optimized for deep learning, similar to cuDNN on NVIDIA GPUs.
Click to reveal the answer
Answer: TrueInfinity Fabric and Multi-GPU Scaling¶
21. Infinity Fabric is a high-speed interconnect technology developed by:
- A. Intel
- B. NVIDIA
- C. AMD
- D. ARM
Click to reveal the answer
Answer: C. AMD22. True or False: Infinity Fabric supports data transfer between GPUs and CPUs, improving multi-GPU performance.
Click to reveal the answer
Answer: TrueGPU Memory Management and AI-Specific Enhancements¶
23. Which of the following is a specialized feature in AMD's RDNA 2 architecture designed for real-time graphics rendering?
- A. Tensor Cores
- B. Ray Accelerators
- C. NVLink
- D. L3 Cache
Click to reveal the answer
Answer: B. Ray Accelerators24. True or False: Matrix Cores in AMD GPUs are specifically designed to accelerate matrix operations in AI workloads.