CUDA Programming/Terminology: Difference between revisions

From CS486wiki
Jump to navigationJump to search
Content deleted Content added
Myuksek1 (talk | contribs)
No edit summary   (change visibility)
Myuksek1 (talk | contribs)
No edit summary   (change visibility)
Line 1: Line 1:
Here are the list of various terms that are specific to GPU Parallel Computing field:
Here are the list of various terms that are specific to GPU Parallel Computing field:


= Terms =
= Kernel =
== CUDA Definition ==

A kernel is a function that, when called, is executed N times in parallel by N different CUDA threads, as opposed to only once like regular C functions.
== Kernel ==
=== CUDA Definition ===
== OpenCL Definition ==
A kernel is a function that, when called, is executed N times in parallel by N different CUDA threads, as opposed to only once like regular C functions.[1]
=== OpenCL Definition ===
A kernel is a function declared in a program and executed on an OpenCL device.[2]
A kernel is a function declared in a program and executed on an OpenCL device.[2]


<i>See also:</i> [[#Kernel_Object|Kernel Object]] definition
<i>See also:</i> [[#Kernel_Object|Kernel Object]] definition


== Kernel Object ==
= Kernel Object =
=== OpenCL Definition ===
== OpenCL Definition ==
A kernel object encapsulates a specific [[#Kernel|kernel]] function declared in a program and the argument values to be used when executing this function. [2]
A kernel object encapsulates a specific [[#Kernel|kernel]] function declared in a program and the argument values to be used when executing this function.


== Compute Capability ==
= Compute Capability =
=== CUDA Definition ===
== CUDA Definition ==
It is defined by a major and a minor revision number. Devices with the same major revision number are of the same core architecture. The minor revision number corresponds to an incremental improvement to the core architecture, possibly including new features.
It is defined by a major and a minor revision number. Devices with the same major revision number are of the same core architecture. The minor revision number corresponds to an incremental improvement to the core architecture, possibly including new features.


= References =
= SIMT =
== CUDA Definition ==
[1] NVIDIA CUDA Computing Guide Version 3.1
SIMT (Single-Instruction, Multiple-Thread) is an architecture that manages the execution of multiple [[#Thread|threads]] concurrently.
[2] The OpenCL Specification, Version 1.1, Revision 36

<i>See also:</i> [[#Warp|Warp]]

= Thread =
== CUDA Definition ==
CUDA thread defines a ready-for-execution/running instance of a [[#Kernel|kernel]]. Each thread has its own instruction address counter and register state.

<i>See also:</i> [[#SIMT|SIMT]]

= Warp =
== CUDA Definition ==
A warp defines a group of 32 parallel threads. Individual [[#Thread|threads]] composing a warp start together at the same program address.

<i>See also:</i> [[#SIMT|SIMT]]

Revision as of 01:50, 21 May 2011

Here are the list of various terms that are specific to GPU Parallel Computing field:

Kernel

CUDA Definition

A kernel is a function that, when called, is executed N times in parallel by N different CUDA threads, as opposed to only once like regular C functions.

OpenCL Definition

A kernel is a function declared in a program and executed on an OpenCL device.[2]

See also: Kernel Object definition

Kernel Object

OpenCL Definition

A kernel object encapsulates a specific kernel function declared in a program and the argument values to be used when executing this function.

Compute Capability

CUDA Definition

It is defined by a major and a minor revision number. Devices with the same major revision number are of the same core architecture. The minor revision number corresponds to an incremental improvement to the core architecture, possibly including new features.

SIMT

CUDA Definition

SIMT (Single-Instruction, Multiple-Thread) is an architecture that manages the execution of multiple threads concurrently.

See also: Warp

Thread

CUDA Definition

CUDA thread defines a ready-for-execution/running instance of a kernel. Each thread has its own instruction address counter and register state.

See also: SIMT

Warp

CUDA Definition

A warp defines a group of 32 parallel threads. Individual threads composing a warp start together at the same program address.

See also: SIMT