Jeff Liu Lab
Home
Projects
Workshop
AI Wiki
AI Lab
Shop
中
Sign In
All
Computing Science
Artificial Intelligence
Deep Learning
Reinforcement Learning
AI Agents
Embodied Intelligence
Robot Engineering
Human-Like Intelligence
AI Engineering
← Back to Wiki
Robot Engineering
Computer Engineering
Computer Engineering Overview
CPU Architecture
GPU & Parallel Computing
Memory & Storage
Bus & Interfaces
Embedded Systems
Operating System Fundamentals
Cameras & Vision
LiDAR
Inertial Navigation & Localization
Motors & Actuators
Communication & Bus
Power Management
Mechanical Structure
Displays & HMI
Force & Tactile Sensing
System Integration
Comments (0)
Sign in to comment
Table of Contents
Overview
GPU Architecture Fundamentals
CPU vs. GPU: Design Philosophy
NVIDIA GPU Architecture
SIMT Execution Model
CUDA Programming Fundamentals
Thread Hierarchy
Memory Hierarchy
CUDA Kernel Example
Image Processing Example
Parallel Speedup Theory
Theoretical Speedup
Gustafson's Law
TensorRT Inference Acceleration
TensorRT Optimization Pipeline
Key Optimization Techniques
Inference Performance Examples
Jetson GPU Specifications Comparison
GPU Programming Best Practices
1. Maximize Occupancy
2. Coalesced Memory Access
3. Minimize CPU-GPU Data Transfers
4. Leverage Tensor Cores
Robot Vision Processing Pipeline
Summary
References
Comments
Comments (0)
Sign in to comment