As multi-core architectures begin to emerge in every area of computing, operating system scheduling that takes the peculiarities of such architectures into account will become mandatory. Due…Full description
As multi-core architectures begin to emerge in every area of computing, operating system scheduling that takes the peculiarities of such architectures into account will become mandatory. Due to arc...
C and C++ are designed basically for making the machine understand the end users (the people uses the computer) instructions. These are classified as high-level programming language which was originally developed by Dennis M.
Leçon d'architectures pour les novicesDescription complète
Engineering college notes
Giuliana Bruno Arqutecturas corporales.Descrição completa
Full description
Esp 8266 and arduino programming
A short collection of programming problems and exercises. Please make comments and suggestions.
Descrição completa
Esp 8266 and arduino programmingFull description
INTERGER AND GOAL PROGRAMMING
The presentation describes data center solutions from major networking vendors. In-depth analysis of individual solutions as well as more technical details are available in the Data Center Fabri...
this subject helps every students to enhance their skills through engaging different kinds of sports activities
edexcel bio u3 practical workFull description
Descripción: Things you can't find in a dictionary
Full description
Full description
Formula
Primauté du visuel et nécessité de moyens de représentation sonores Thèse de Doctorat. École d’ArchitectureDescription complète
n
VALLIAMMAI ENGINEERING COLLEGE SRM Nagar, Kattankulathur – 603 203
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING QUESTION BANK
VIII SEMESTER CS6801- Multi Core Architectures and Programming Regulation – 2013 Academic Year 2016 – 17 Prepared by Mr.L.Karthikeyan Assistant Professor /CSE Mr.M.Mayuranathan, Assistant Professor/CSE
n
VALLIAMMAI ENGNIEERING COLLEGE SRM Nagar, Kattankulathur – 603203. DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING
Year & Semester Subject Code Subject Name Degree & Branch Staff in charge
S.No
: : : : :
IV Year/VII Semester CS6801 Multi-Core Architecture and programming M.E CSE Mr.M.Mayuranathan and Mr.L.Karthiekeyan
QUESTIONS UNIT I MULTI-CORE PROCESSORS
COMPETENCE
LEVEL
SYLLABUS Single core to Multi-core architectures – SIMD and MIMD systems – Interconnection networks Symmetric and Distributed Shared Memory Architectures – Cache coherence - Performance Issues – Parallel program design. 1.
Define Vector Instruction.
Remember
BTL 1
2.
BTL 5
4.
Evaluate Compare between symmetric memory Architecture and distributed Memory Architecture Generalize are the factors to increasing the operating frequency of Create the processor? Discuss the issues available in handling the performance? Understand
5.
Show and define SIMD System
Apply
BTL3
6.
compose the definition of MIMD System
Create
BTL6
7.
Express NUMA with neat sketch
Understand
BTL2
8.
Pointout Interconnection networks and its types
Analyse
BTL4
9.
Pointout Toroidal mesh with neat diagram
Analyse
BTL4
10.
Give definition of latency and bandwidth.
Understand
BTL2
11.
Show the neat diagram for structural model of centralized shared Apply memory multiprocessor
3.
BTL 6 BTL2
BTL3
n 12.
Define Cache Coherence protocol with its types
Remember
BTL1
13.
Define Directory based.
Remember
BTL1
14.
Pointout Snooping.
Analyse
BTL4
15.
Give the characteristic of the performance due to write update and
Understand
BTL2
write invalidate protocol? 16.
Tell the disadvantage of Symmetric shared memory architecturte?
Remember
BTL1
17.
Define Agglomeration or aggregation.
Remember
BTL1
18.
Compare singe and multi core CPU
Evaluate
BTL5
19.
Define false sharing.
Remember
BTL1
20.
Apply
BTL3
Understand
BTL2
Understand
BTL2
3
Show a mathematical formula for speedup and efficiency of parallel program speed up. PART- B i) Summarize the Evaluation from single core to multi core architecture ii) Compare single core and multi core processor Discuss SIMD . Describe MIMD
Remember
BTL1
4
Illustrate Discuss Shared Memory Interconnect.
Apply
BTL3
5
Explain Distributed Memory Interconnect.
Analyze
BTL4
6
Explain Symmetric shared Memory Architecture.
Analyze
BTL4
7
Discuss Distributed Shared Memory Architecture.
Understand
BTL2
8
Illustrate Cache Coherence.
Apply
BTL3
9
Identify the Performance issues and discuss briefly.
Remember
BTL1
10
Explain Directory based cache coherence protocol
Analyze
BTL4
11
Examine the Amdahl Law.
Remember
BTL1
12
Describe the Serial program.
Remember
BTL1
13
Generalize the Snooping protocol briefly.
Create
BTL6
1
2
n 14
Summarize the Parallelizing the serial program
Evaluate
BTL5
UNIT II PARALLEL PROGRAM CHALLENGES SYLLABUS Performance – Scalability – Synchronization and data sharing – Data races – Synchronization primitives (mutexes, locks, semaphores, barriers) – deadlocks and livelocks – communication between threads (condition variables, signals,message queues and pipes). 1
Define Algorithm complexity?
Remember
BTL1
2
Tell the Out-of-order in detail?
Remember
BTL1
3
Pointout Hardware prefetching
Analyze
BTL4
4
Give the Definition of Software prefetching.
Understand
BTL2
5
Define Work imbalance.
Remember
BTL1
6
Generalize hot or contended locks.
Create
BTL6
7
Give the define Oversubscription
Understand
BTL2
8
Tell the processes of priority inversion.
Create
BTL6
9
Assess the Definition Data races.
Evaluate
BTL5
10
Assess the Definition Synchronization
Evaluate
BTL5
11
Define region of code.
Remember
BTL1
12
Give the definition deadlock
Understand
BTL2
13
Define livelock
Remember
BTL1
14
Examine the condition variable.
Apply
BTL3
15
Complete the definition Signal.
Apply
BTL3
16
Pointout Event.
Analyze
BTL4
17
Examine the Message queue.
Apply
BTL3
n
18
Pointout Named pipes
Analyze
BTL4
19 20
Give the Example two threads in deadlock. Define mutex,critical region and semaphores PART- B Discuss in details about the importance of algorithmic complexity Explain the Outline about necessity of structure reflects in performance
Understand Remember
BTL2 BTL1
Understand
BTL2
Analyze
BTL4
Remember
BTL1
Remember Understand
BTL1 BTL2
1 2 3
Describe briefly about Data Structures plays vital role in part of performance with an example.
4 5
Describe briefly the Super linear scaling. Give an outline about scaling of library code
6
Write in detail and summarize about Hardware constraints applicable Evaluate to improve scaling.
BTL5
7 8
Apply Illustrate about operating system constraints to scaling. Give briefly the List in details about tools used for detecting data Apply races. Create Generalize in detail about Data races and to overcome it.
BTL3 BTL2
Describe in detail the Mutex, and semaphore Explain briefly the Spin lock and reader and writer lock. Describe in detail about conditional variable in communication between thread. Explain and Write short note signal ,events and message queue Illustrate and explain the Named pipes
Remember Analyze Remember
BTL1 BTL4 BTL1
Analyze Apply
BTL4 BTL3
9 10 11 12 13 14
BTL6
UNIT III SHARED MEMORY PROGRAMMING WITH OpenMP SYLLABUS OpenMP Execution Model – Memory Model – OpenMP Directives – Work-sharing Constructs – Library functions –Handling Data and Functional Parallelism – Handling Loops – Performance Considerations. 1 2 3 4 5
Compose the statement termed as initial task region? Pointout the List the effect of cancel construct. When will a cancellation point construct triggers? Define the term thread private memory. Differentiate between shared memory model and message passive model.
Create Analyze Remember Remember Understand
BTL6 BTL4 BTL1 BTL1 BTL2
n
6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 1 2 3
4 5
6 7 8
9 10 11 12 13 14
Give short note on private variable. Pointout about flush-set Express how to enable consistency between two threads temporary view and memory. Order the criteria for sentinel of condition compilation.
Understand Analyze Understand
BTL2 BTL4 BTL2
Analyze
BTL4
Examine are the ICVs stored values affect loop region? Identify the ICVs stored values affect program execution. List the restriction to array. Give the List of the restrictions to parallel construct. Show the List of the restrictions to work sharing constructs. Show the List of the restriction to sections constructs. Describe about simple lock routines. Examine the about Nestable lock routines Assess about pragma. Compose the definition the term shared variable in execution context. Conclude the run-time system know how many threads to create PART- B Generalize briefly about OPENMP execution model Explain the types Shared memory model
i) Examine the Lock routines(8) ii) Examine the Portable timer routines(8) Explain briefly about the General Data parallelism.
Remember
BTL1
Analyze
BTL4
Describe briefly about the Functional parallelism. Describe briefly about the Handling loops. Illustrate briefly about the Performance considerations i) Explain the Nowait clause(8) ii) Explain the Single Pragma(8)
Understand Understand Apply Analyze
BTL2 BTL2 BTL3 BTL4
i) Discuss Directive formate? (8) ii) Discuss conditional compilation(8) Collect the all information about Internal control variables(16) i) ii) iii) i) ii) i) ii) i) ii)
Explain the Array section(4) Explain the Canonical loop form(6) Explain the Parallel construct(6) Describe the loop constrct(8) Describe the sections consructs(8) Discuss single construct(8) Discuss workshare construct(8) Illustrate the runtime library definitions(8) Illustrate the Execution environment routines(8)
n
UNIT IV DISTRIBUTED MEMORY PROGRAMMING WITH MPI SYLLABUS MPI program execution – MPI constructs – libraries – MPI send and receive – Point-to-point andCollective communication – MPI derived datatypes – Performance evaluation. Define the term MPI Give the term collective communications What is the purpose of wrapper script? Tell, How to compile and execute a mpi_hello.c program in mpi environment? What are the functions in MPI to initiate and terminates a computation, identify processes, and send and receive message. List the different datatype constructors Complete about communicators in MPI Pointout to remove a group? List the dimension of the array in distributed array. Complete, How to create a Cartesian constructor? Complete, What are the interfaces to create a distributed graph topology? List the features of blocking and non-blocking in point-to-point communication?
Remember Understand Create Remember
BTL1 BTL2 BTL6 BTL1
Remember
BTL1
Analyze Apply Analyze Remember Apply Apply
BTL4 BTL3 BTL4 BTL1 BTL3 BTL3
Remember
BTL1
13 14 15 16 17 18 19
List some predefined reduction operators in MPI Examine about MPI_All_reduce and their representations Pointout the term broadcast in collective communication Assess, List the functions of Group Accessors. How to plan to represent any collection of data items in MPI? Predict, How will calculate elapsed time in MPI? Express the term linear speedup
PART- B i) Describe about MPI Program execution(8) ii) Describe about MPI Init and Finalize(8) Explain briefly about the MPI Program. i) Describe about the Datatype constructors(8) ii) Discuss about the Subarray Datatype constructor(8) i) Explain the Distributed array datatype constructor(6) ii) Explain the Cartesian constructor(6) iii) Explain the Distributed graph constructor(4) i) Generalize the Group of process(8) ii) Explain the Virtual topology(8) i) Describe the Attribute Caching(8) ii) Discuss about the Communicators(8)
n
7 8 9 10 11
12 13 14
Apply Illustrate and explain about the MPI Send. Describe about the MPI Receive. Remember Remember Describe about the Point-to-Point communication Discuss about the Collective communication Understand i) Demonstrate the Compare point-to-point communication Apply and collective communication. ii) Discuss about the Scatter and gather Explain about the MPI derived datatypes Analyze Identify the Performance evaluation in detail. Remember Explain about the Tree-structured communication Analyze UNIT V PARALLEL PROGRAM DEVELOPMENT 9
BTL3 BTL1 BTL1 BTL2 BTL3
BTL4 BTL 1 BTL4
SYLLABUS Case studies - n-Body solvers – Tree Search – Open MP and MPI implementations and comparison. 1
Tell the input and output parameters of n-body problem
Remember
BTL1
2
Tell about reduced algorithm.
Remember
BTL1
3
Pointout about Pthread’s Loop_schedule functionality in
Analyze
BTL4
parallel processing. 4 5
Discuss the different data structures can be adopted in a process? Tell the two phases for computation of forces?
Understand Remember
BTL2 BTL1
6 7 8 9 10
Pointout the term ring pass. Demonstrate graph? Examine the directed graph. Explain, Why digraph is used in Travelling salesperson problem? Show, How to find a least cost tour in TSP
Analyze Apply Apply Evaluate Apply
BTL4 BTL3 BTL3 BTL5 BTL3
11
Give, how to function push_copy is used in TSP
Understand
BTL2
12 13
Tell the global variables for recursive DFS? Define the term Pthreads or POSIX Thread
Remember Remember
BTL1 BTL1
14
Formulate the statement, the different categories of Pthread?
Create
BTL6
15
Analyze
BTL4
Remember
BTL1
17 18 19
Pointout are the reasons for parameter thread-in_cond_wait used in Tree search? Tell the modes od message passing interfaces for send and its functions? Discuss about My_avail_tour_count functions Discuss about Pthread_mutex_trylock Write about fulfill_request functions.
Understand Evaluate Create
BTL2 BTL5 BTL6
20
Distinguish between MPI_pack and MPI_unpack
Understand
BTL2
16
n
1
Generalize about the Two serial programs.
Create
BTL6
2 3 4 5 6
Explain about the Parallelizing the n-body solvers. Explain about the Parallelizing the basic solver using OpenMP Describe about the Parallelizing the reduced solver using OpenMP Explain about the Parallelizing the basic solver using MPI Explain about the Parallelizing the reduced solver using MPI
Analyze Evaluate Understand Analyze Analyze
BTL4 BTL5 BTL2 BTL4 BTL4
7 8
Examine the Performance of the MPI solvers. Express detail about Recursive DFS
Apply Understand
BTL3 BTL2
9 10 11 12
Express detail about Non recursive DFS Examine the Data structure for the serial Implementation
Remember Apply Express detail about the Static parallelizing of tree search using Pthreads Remember i) Describe the dynamic parallelizing of tree search using Understand
BTL1 BTL3 BTL1 BTL2
ii)
13
14
i) ii) iii)
Pthreads(8) Describe the termination(8)
Describe the Evaluating the Pthread Tree(8) Remember Describe the search Program(4) Describe the Parallelizing the tree-search program using OpenMP(4) Describe and Compare OpenMP and MPI implementation Remember