High Performance Computing (HPC)/High Performance Computing (HPC) MCQ Set 2 Sample Test,Sample questions

Question:
 A good mapping does not depends on which following factor

1. knowledge of task sizes

2.the size of data associated with tasks

3.characteristics of inter-task interactions

4.task overhead


Question:
 All nodes collects _____ message corresponding to √p nodes to their respectively

1. √p

2.p

3. p+1

4. p-1


Question:
 All-to-all personalized communication can be used in ____

1.fourier transform

2.matrix transpose

3.sample sort

4.All of the above


Question:
 All-to-one communication (reduction) is the dual of ______ broadcast.

1.all-to-all

2. one-to-all

3.one-to-one

4.all-to-one


Question:
 Cost Analysis on a mesh is

1. A. 2ts(sqrt(p) + 1) + twm(p - 1)

2.2tw(sqrt(p) + 1) + tsm(p - 1)

3.2tw(sqrt(p) - 1) + tsm(p - 1)

4.2ts(sqrt(p) - 1) + twm(p - 1)


Question:
 Goal of good algorithm is to implement commonly used _____ pattern.

1. communication

2.interaction

3.parallel

4.regular


Question:
 If we port algorithm to higher dimemsional network it would cause

1.error

2.contention

3.recursion

4.none


Question:
 In ------------task are defined before starting the execution of the algorithmting?

1. dynamic task

2.static task

3.regular task

4.one way task


Question:
 In a broadcast and reduction on a balanced binary tree reduction is done in ______

1.recursive order

2. straight order

3.vertical order

4.parallel order


Question:
 In one -to- all broadcast there is

1. divide and conquer type algorithm

2.sorting type algorithm

3.searching type algorithm

4. simple algorithm


Question:
 In Scatter Operation on Hypercube, on each step, the size of the messages communicated is ____

1. tripled

2.halved

3.doubled

4.no change


Question:
 In the scatter operation ____ node send message to every other node

1.single

2.double

3.triple

4.none


Question:
 In the second phase of 2D Mesh All to All, the message size is ___

1.m

2.p*sqrt(m)

3.p

4.m*sqrt(p)


Question:
 In this decomposition problem decomposition goes hand in hand with its execution

1.data decomposition

2. recursive decomposition

3.explorative decomposition

4.speculative decomposition


Question:
 It is not possible to port ____ for higher dimensional network

1.algorithm

2.hypercube

3.both

4.none


Question:
 Reduction can be used to find the sum, product, maximum, minimum of _____ of numbers.

1.tuple

2.list

3.sets

4.all of above


Question:
 subsets of processes in ______ interaction.

1.global

2.local

3.wide

4.variable


Question:
 threads being block altogether and being executed in the sets of 32 threads called a

1. thread block

2.32 thread

3.32 block

4.unit block


Question:
 When the topological sort of a graph is unique?

1.when there exists a hamiltonian path in the graph

2.in the presence of multiple nodes with indegree 0

3.in the presence of single node with indegree 0

4.in the presence of single node with outdegree 0


Question:
 Which of the following is not a form of parallelism supported by CUDA

1.vector parallelism - floating point computations are executed in parallel on wide vector units

2. thread level task parallelism - different threads execute a different tasks

3.block and grid level parallelism - different blocks or grids execute different tasks

4.data parallelism - different threads and blocks process different parts of data in memory


Question:
 which of the following is not a granularity type

1.course grain

2.large grain

3.medium grain

4.fine grain


Question:
 which of the following is not an example of explorative decomposition

1.n queens problem

2.15 puzzal problem

3. tic tac toe

4.quick sort


Question:
accumulate results and send with the same pattern is...

1.broadcast

2.naive approach

3.recursive doubling

4.reduction symmetric


Question:
All processes participate in a single ______ interaction operation.

1.global

2.local

3.wide

4.variable


Question:
all processes that have the data can send it again is

1.recursive doubling

2. naive approach

3.reduction

4.All of the above


Question:
All-to-all broadcast algorithm for the 2D mesh is based on the

1. linear array algorithm

2. ring algorithm

3.both

4.none


Question:
blocking optimization is used to improve temmporal locality for reduce

1. hit miss

2.misses

3.hit rate

4.cache misses


Question:
Broadcast and reduction operations on a mesh is performed

1. along the rows

2.along the columns

3.both a and b concurrently

4.none of these


Question:
Communication between two directly link nodes

1.cut-through routing

2.store-and-forward routing

3.nearest neighbour communication

4.none


Question:
Cost Analysis on a ring is

1. (ts + twm)(p - 1)

2. (ts - twm)(p + 1)

3. (tw + tsm)(p - 1)

4.(tw - tsm)(p + 1)


Question:
CUDA thought that 'unifying theme' of every form of parallelism is

1.cda thread

2.pta thread

3.cuda thread

4.cud thread


Question:
Each node first sends to one of its neighbours the data it need to....

1. broadcast

2.identify

3.verify

4.none


Question:
efficiency of data parallel algorithm depends on the

1.efficient implementation of the algorithm

2.efficient implementation of the operation

3.both

4.none


Question:
every node has to know when to communicate that is

1.call the procedure

2.call for broadcast

3. call for communication

4.call the congestion


Question:
every node on the linear array has the data and broadcast on the columns with the linear array algorithm in _____

1. parallel

2.vertical

3.horizontal

4.all


Question:
For sake of simplicity, the number of nodes is a power of

1.1

2.2

3.3

4.4


Question:
Generalization of broadcast in Which each processor is

1.source as well as destination

2.only source

3.only destination

4.none


Question:
Group communication operations are built using which primitives?

1.one to all

2.all to all

3.point to point

4.none of these


Question:
if "X" is the message to broadcast it initially resides at the source node

1.1

2.2

3.8

4.0


Question:
In a balanced binary tree processing nodes is equal to

1.. leaves

2. number of elemnts

3.branch

4.none


Question:
In a eight node ring, node ____ is source of broadcast

1.1

2.2

3.8

4.0


Question:
In All to All on Hypercube, The size of the message to be transmitted at the next step is ____ by concatenating the received message with their current data

1.doubled

2.tripled

3.halfed

4.no change


Question:
In collective communication operations, collective means

1.involve group of processors

2.involve group of algorithms

3. involve group of variables

4.none of these


Question:
In task dependency graph longest directed path between any pair of start and finish node is called as --------------

1.. total work

2. critical path

3.task path

4.task path


Question:
In the first phase of 2D Mesh All to All, the message size is ___

1. p

2.m*sqrt(p)

3.m

4.p*sqrt(m) discuss


Question:
kernel may contain only host code

1.a code known as grid which runs on GPU consisting of a set of A. 32 thread

2.unit block

3.32 block

4. thread block


Question:
logical operators used in algorithm are

1. xor

2.and

3.both

4.none


Question:
Nides with zero in i least significant bits participate in _______

1.algorithm

2.broadcast

3.communication

4.searching


Question:
one to all broadcast use

1.recursive doubling

2. simple algorithm

3.both

4.none


Question:
One-to-All Personalized Communication operation is commonly called ___

1.gather operation

2.concatenation

3. scatter operation

4.none


Question:
only connections between single pairs of nodes are used at a time is

1. good utilization

2.poor utilization

3. massive utilization

4.medium utilization


Question:
Renaming relative to the source is _____ the source.

1. xor

2.xnor

3.and

4.nand


Question:
Renaming relative to the source is _____ the source.

1.xor

2.xnor

3.and

4.nand


Question:
Similar communication pattern to all-to-all broadcast except in the_____

1. reverse order

2. parallel order

3.straight order

4.vertical order


Question:
source ____ is bottleneck.

1.process

2.algorithm

3.list

4.tuple


Question:
Task dependency graph is ------------------

1.directed

2.undirected

3.directed acyclic

4.undirected acyclic


Question:
The algorithm terminates in _____ steps

1.p

2.p+1

3.p+2

4. p-1


Question:
The all-to-all broadcast on Hypercube needs ____ steps

1.p

2.sqrt(p) - 1

3.log p

4.none


Question:
The dual of all-to-all broadcast is

1.all-to-all reduction

2.all-to-one reduction

3.both

4.none


Question:
The dual of the scatter operation is the

1.concatenation

2.gather operation

3.both

4.none


Question:
The gather Operation is exactly the inverse of _____

1.scatter operation

2. recursion operation

3.execution

4.none


Question:
the procedure is disturbed and require only point-to-point _______

1.synchronization

2.communication

3.both

4.none


Question:
the procedure is disturbed and require only point-to-point _______

1.synchronization

2.communication

3.both

4.none


Question:
The processors compute ______ product of the vector element and the loval matrix

1.local

2.global

3.both

4.none


Question:
The second communication phase is a columnwise ______ broadcast of consolidated

1.all-to-all

2.one -to-all

3.all-to-one

4.point-to-point


Question:
The style of parallelism supported on GPUs is best described as

1.misd - multiple instruction single data

2.simt - single instruction multiple thread

3.sisd - single instruction single data

4.mimd


Question:
The ____ do not snoop the messages going through them.

1. nodes

2.variables

3.tuple

4.list


Question:
using different links every time and forwarding in parallel again is

1. better for congestion

2.better for reduction

3.better for communication

4.better for algorithm


Question:
What is a high performance multi-core processor that can be used to accelerate a wide variety of applications using parallel computing.

1.cpu

2.dsp

3.gpu

4.clu


Question:
Which is also called "Total Exchange" ?

1.all-to-all broadcast

2.all-to-all personalized communication

3.all-to-one reduction

4.none


Question:
Which is known as Broadcast?

1.one-to-one

2.one-to-all

3.all-to-all

4.all-to-one


Question:
Which is known as Reduction?

1.. all-to-one

2.all-to-all

3. one-to-one

4.one-to-all


Question:
Which of the following correctly describes a GPU kernel

1. a kernel may contain a mix of host and gpu code

2.all thread blocks involved in the same computation use the same kernel

3. a kernel is part of the gpus internal micro-operating system, allowing it to act as in independent host

4.kernel may contain only host code


Question:
which of the following is a an example of data decomposition

1.matrix multiplication

2. merge sort

3.quick sort

4.15 puzzal


Question:
which of the following is not a granularity type

1. course grain

2. large grain

3. medium grain

4.fine grain


Question:
which of the following is not an parallel algorithm model

1.data parallel model

2.task graph model

3.task model

4.work pool model


Question:
which of the following is not the array distribution method of data partitioning

1.block

2.cyclic

3.block cyclic

4.chunk


Question:
which problems can be handled by recursive decomposition

1.backtracking

2.greedy method

3.divide and conquer problem

4.branch and bound


Question:
wimpleat way to send p-1 messages from source to the other p-1 processors

1.algorithm

2.communication

3.concurrency

4.receiver


Question:
__ can be performed in an identical fashion by inverting the process.

1.recursive doubling

2.reduction

3.broadcast

4.none of these


More MCQS

  1. Green Computing (GC)
  2. High Performance Computing (HPC) MCQ Set 1
  3. High Performance Computing (HPC) MCQ Set 2
Search
Olete Team
Online Exam TestTop Tutorials are Core Java,Hibernate ,Spring,Sturts.The content on Online Exam Testwebsite is done by expert team not only with the help of books but along with the strong professional knowledge in all context like coding,designing, marketing,etc!