Parallel
computation is one of the computer programming that
allows to execute commands simultaneously and concurrently in a single or
multiple processors inside a CPU. Parallel computation itself is useful to
improve the performance of the computer as more and more processes that can be
done at the same time it will be faster.
Parallel
concept
The
concept of parallel is a processors ability to perform a task
or multiple tasks simultaneously or concurrently, in other words, the processor
is able to perform one or many tasks at one time.
Distributed
Processing
Distributed
processing is the process of parallel processing in distributed parallel
processing using multiple machines. So, it could be said the ability of the computers
that run simultaneously to solve a problem with the process quickly.
Architectural Parallel
Computer
According
to a Processor Designer, taxonomy Flynn, Computer Architecture is divided
into four sections.
1. SISD
(Single Instruction Single Data Stream)
The
type of computer that only has one processor and one instruction is executed
serially.
2. SIMD
(Single Instruction Multiple Data Stream)
This type of computer
that has more than one processor, but this computer only executes one instruction
in parallel on different data in lock-step level.
3. MISD
(Multiple Instruction Single Data Stream)
This type of computer
that has one processor and execute multiple instructions in parallel but in
practice there is no computer that is built with this architecture because the system
is not easily understood, until now there has been no computers that use this
type of architecture
4. MIMD
(Multiple Instruction Multiple Data Stream)
This type of computer that has more than one processor and execute more
than one instruction in parallel. This type of computer that is most widely
used to build a parallel computer, even many supercomputer that implement
this architecture, because the
models and concepts that are not too complicated to understand.
Introduction
to Programming Thread
A
thread in computer programming is
a relevant information about the use of a single program that can handle
multiple users simultaneously.Thread This allows the program to determine how
the user entered into the program in turn and the user will go back to using a
different user. Multiple threads can run concurrently with other processes
divides the resources into memory, while the other
processes do not share it.
Introduction to
Programming CUDA GPU
GPU
Refers to a specific processor GPU to accelerate and change the memory to speed
up image processing. The GPU itself is usually located on the graphics card
or laptop computer
CUDA
(Compute Unified Device Architecture)
is a scheme created by NVIDIA as the GPU (Graphic Processing Unit) capable of
computing not only to graphics processing, but also for general purposes. So with
the CUDA we can take advantage of multiple processors from NVIDIA to do the
calculation process much or computing.
REFFERENCE
· http://uchaaii.blogspot.com/2013/07/parallel-computation.html