...

GPU VSIPL: High Performance VSIPL Implementation for GPUs Andrew Kerr, Dan Campbell*,

by user

on
Category: Documents
17

views

Report

Comments

Transcript

GPU VSIPL: High Performance VSIPL Implementation for GPUs Andrew Kerr, Dan Campbell*,
GPU VSIPL: High Performance
VSIPL Implementation for GPUs
Andrew Kerr, Dan Campbell*,
Mark Richards, Mike Davis
[email protected], [email protected],
[email protected], [email protected]
High Performance Embedded Computing (HPEC)
Workshop
24 September 2008
Distribution Statement (A): Approved for
public release; distribution is unlimited
This work was supported in part by DARPA and AFRL
under contracts FA8750-06-1-0012 and FA8650-07-C7724. The opinions expressed are those of the authors.
GTRI_B-‹#›
1
Signal Processing on Graphics Processors
• GPUs original role: turn 3-D polygons into 2-D pixels…
• …Which also makes them cheap & plentiful source of FLOPs
• Leverages volume & competition in entertainment industry
• Primary role highly parallel, very regular
• Typically <$500 drop-in addition to standard PC
• Outstripping CPU capacity, and growing more quickly
• Peak theoretical ~1TFlop
• Power draw: 280GTX = 200W
Q6600 = 100W
• Still making improvements in market app with more
parallelism, so growth continues
GTRI_B-‹#›
2
GPU/CPU Performance Growth
CPU & GPU Capacity Growth
1171
1000
914
GFLOPS
244
122
100
85
60
24
10
13
12
ATI
NVIDIA
Intel x86
"Moore's Law"
GTRI_B-‹#›
3
GPGPU (Old) Concept of Operations
“A=B+C”
void main(float2 tc0 : TEXCOORD0,
out float4 col : COLOR,
uniform samplerRECT B,
uniform samplerRECT C)
{
col = texRECT (B, tc0) +
texRECT (C, tc0);
}
• Arrays  Textures
• Render polygon with the same pixel dimensions as output texture
• Execute with fragment program to perform desired calculation
• Move data from output buffer to desired texture
Now we have compute-centric programming models…
… But they require expertise to fully exploit
GTRI_B-‹#›
4
VSIPL - Vector Signal Image Processing Library
• Portable API for linear algebra, image & signal processing
• Originally sponsored by DARPA in mid ’90s
• Targeted embedded processors – portability primary aim
• Open standard, Forum-based
• Initial API approved April 2000
• Functional coverage
• Vector, Matrix, Tensor
• Basic math operations, linear algebra, solvers, FFT, FIR/IIR,
bookkeeping, etc
GTRI_B-‹#›
5
VSIPL & GPU: Well Matched
• VSIPL is great for exploiting GPUs
• High level API with good coverage for dense linear algebra
• Allows non experts to benefit from hero programmers
• Explicit memory access controls
• API precision flexibility
• GPUs are great for VSIPL
• Improves prototyping by speeding algorithm testing
• Cheap addition allows more engineers access to HPC
• Large speedups without needing explicit parallelism at application level
GTRI_B-‹#›
6
GPU-VSIPL Implementation
• Full, compliant implementation of VSIPL Core-Lite Profile
• Fully encapsulated CUDA backend
• Leverages CUFFT library
• All VSIPL functions accelerated
• Core Lite Profile:
• Single precision floating point, some basic integer
• Vector & Sxalar, complex & real support
• Basic elementwise, FFT, FIR, histogram, RNG, support
• Full list: http://www.vsipl.org/coreliteprofile.pdf
• Also, some matrix support, including vsip_fftm_f
GTRI_B-‹#›
7
CUDA Programming & Optimization
CUDA Programming Model
CUDA Optimization Considerations
Grid
Block
Block
Block
Datapath
Datapath
• Maximize occupancy to hide
Datapath
Datapath
Shared
Shared
Datapath
Memory
Datapath
Thread
Memory
Thread
Datapath
Shared
Datapath
Thread
Datapath
Register
Memory
Datapath
Register
Thread
File
Datapath
File
Datapath
Datapath
Datapath
Datapath
Datapath
Register
File
•
•
•
•
•
•
•
memory latency
Keep lots of threads in flight
Carefully manage memory access to
allow coalesce & avoid conflicts
Avoid slow operations (e.g integer
multiply for indexing)
Minimize synch barriers
Careful loop unrolling
Hoist loop invariants
Reduce register use for greater
occupancy
Device Memory
CPU
Dispatch
• “GPU Performance Assessment with
the HPEC Challenge” – Thursday PM
Host Memory
GTRI_B-‹#›
8
GPU VSIPL Speedup: Unary
320X
1000
nVidia 8800GTX
Vs
Intel Q6600
Speedup
100
10
20X
1
0.1
Vector Length
vcos
vexp
vlog
vsqrt
vmag
vsq
cvconj
GTRI_B-‹#›
9
GPU VSIPL Speedup: Binary
100
40X
nVidia 8800GTX
Vs
Intel Q6600
Speedup
10
25X
1
0.1
0.01
Vector Length
vmul
vadd
vdiv
vsub
cvadd
cvsub
cvmul
cvjmul
GTRI_B-‹#›
10
GPU VSIPL Speedup: FFT
83X
100
nVidia 8800GTX
Vs
Intel Q6600
10
Speedup
39X
1
0.1
0.01
Vector Length
Real to Complex
Complex to Complex
GTRI_B-‹#›
11
GPU VSIPL Speedup: FIR
Filter Length 1024
10000
Runtime(ms)/Speedup
1000
nVidia 8800GTX
Vs
Intel Q6600
157X
100
10
1
32768
65536
131072
262144
524288
1048576
2097152
Input Vector Length
GPU (r)
GPU (c)
TASP (r)
TASP (c)
Speedup (r)
Speedup (c)
GTRI_B-‹#›
12
Application Example: Range Doppler Map
• Simple Range/Doppler data visualization application demo
• Intro app for new VSIPL programmer
• 59x Speedup TASP  GPU-VSIPL
• No changes to source code
Section
Admit
Baseband
Zeropad
Fast time FFT
Multiply
Fast Time FFT-1
Slow time FFT, 2x CT
log10 |.|2
Release
Total:
9800GX2
Time (ms)
8.88
67.77
23.18
47.25
8.11
48.59
12.89
22.2
54.65
293.52
Q6600
Time (ms)
0
1872.3
110.71
5696.3
33.92
5729.04
3387
470.15
0
17299.42
Speedup
0
28
5
121
4
118
263
21
0
59
GTRI_B-‹#›
13
GPU-VSIPL: Future Plans
• Expand matrix support
• Move toward full Core Profile
• More linear algebra/solvers
• VSIPL++
• Double precision support
GTRI_B-‹#›
14
Conclusions
• GPUs are fast, cheap signal processors
• VSIPL is a portable, intuitive means to exploit GPUs
• GPU-VSIPL allows easy access to GPU performance without
becoming an expert CUDA/GPU programer
• 10-100x speed improvement possible with no code change
• Not yet released, but unsupported previews may show up at:
http://gpu-vsipl.gtri.gatech.edu
GTRI_B-‹#›
15
Fly UP