This course aims at providing the foundations about computing as a collection of tasks that may be executing simultaneously and potentially interacting with each other. These tasks can be executed on a single or multiple processors or distributed across a network. The course presents the models, challenges, algorithms and systems focusing on three main aspects/modules: concurrency (multiple computations interacting with each other), parallelism (multiple cores or processors), and distribution (multiple computers across a network).
Following a set of introductory sessions, the course has three modules: concurrency, parallelism and distribution. The student has to select two of the three modules. The lectures are complemented with programming exercices to illustrate the problems and evaluate the solutions.
Teachers
Person in charge
Jordi Guitart Fernandez (
)
Others
Jorge Castro Rabal (
)
Josep Ramon Herrero Zaragoza (
)
Weekly hours
Theory
2
Problems
0
Laboratory
2
Guided learning
0.5
Autonomous learning
8
Competences
Technical Competences of each Specialization
Computer networks and distributed systems
CEE2.1 - Capability to understand models, problems and algorithms related to distributed systems, and to design and evaluate algorithms and systems that process the distribution problems and provide distributed services.
CEE2.3 - Capability to understand models, problems and mathematical tools to analyze, design and evaluate computer networks and distributed systems.
High performance computing
CEE4.2 - Capability to analyze, evaluate, design and optimize software considering the architecture and to propose new optimization techniques.
Generic Technical Competences
Generic
CG1 - Capability to apply the scientific method to study and analyse of phenomena and systems in any area of Computer Science, and in the conception, design and implementation of innovative and original solutions.
CG5 - Capability to apply innovative solutions and make progress in the knowledge to exploit the new paradigms of computing, particularly in distributed environments.
Transversal Competences
Teamwork
CTR3 - Capacity of being able to work as a team member, either as a regular member or performing directive activities, in order to help the development of projects in a pragmatic manner and with sense of responsibility; capability to take into account the available resources.
Reasoning
CTR6 - Capacity for critical, logical and mathematical reasoning. Capability to solve problems in their area of study. Capacity for abstraction: the capability to create and use models that reflect real situations. Capability to design and implement simple experiments, and analyze and interpret their results. Capacity for analysis, synthesis and evaluation.
Objectives
Understand the definition of distributed system and its possible applications, as well as challenges to be faced to design and implement it.
Related competences:
CEE2.1,
CG5,
Subcompetences:
Know the possible applications of a distributed system
Understand the challenges to design and implement a distributed system: heterogeneity, lack of global view, concurrency, lack of a single point of control, security, asynchrony, openness, transparency, fault tolerance, scalability
Put examples of distributed systems
Understand the definition of distributed system
Understand the problem of time and events ordering in a distributed system and explain and implement the mechanisms of logic clocks to attack this problem and algorithms to synchronize physical clocks in a distributed system.
Related competences:
CEE2.3,
CEE2.1,
Subcompetences:
Understand the problem of time and events ordering in a distributed system
Explain and implement logical clocks mechanisms to attack this problem: happened-before relation, Lamport logical clocks (scalar, vector)
Explain and implement algorithms to synchronize physical clocks in a distributed system: Cristian (NTP), Berkeley
Understand the problem of obtaining a consistent global state in a distributed system and explain the distributed snapshot mechanism to attack this problem, as well as define predicates for global assessment of properties in a distributed system.
Related competences:
CEE2.3,
CEE2.1,
Subcompetences:
Understand the problem of obtaining a consistent global state in a distributed system
Define global predicates for evaluating properties in a distributed system: properties of predicates (stability), occurrence of predicates (possibly, definitely)
Explain the Chandy and Lamport's mechanism of distributed snapshot to attack this problem
Describe, compare and implement algorithms for the coordination of processes in a distributed system, including the coordination necessary for leader election, multicast group communication and consensus.
Related competences:
CEE2.3,
CEE2.1,
CTR3,
Subcompetences:
Describe, compare, and implement algorithms for the coordination of processes in a distributed system for multicast group communication: basic reliable multicast, scalable reliable multicast, ordered multicast (FIFO, causal, total), atomic multicast
Describe, compare, and implement algorithms for the coordination of processes in a distributed system to ensure consensus: the two army problem, Dolev & Strong's algorithm, the Byzantine generals problem, Paxos
Describe, compare, and implement algorithms for the coordination of processes in a distributed system for the election of leader: Bully, Chang & Roberts, Ring
Understand the problem of concurrent execution of transactions and describe, compare, and implement different concurrency control mechanisms.
Related competences:
CEE2.3,
CEE2.1,
CTR6,
Subcompetences:
Describe, compare, and implement different concurrency control mechanisms: two-phase locking (including deadlock detection and treatment), optimistic concurrency control, timestamp ordering
Understand the problem of concurrent execution of transactions: lost update, inconsistent retrievals, serial equivalence, recovery of aborts (dirty read, write premature)
Extend the concept of transaction, the commit protocol, and the concurrency control mechanisms in a distributed system.
Related competences:
CEE2.3,
CEE2.1,
Subcompetences:
Extend concurrency control mechanisms in a distributed system: two-phase locking (including distributed deadlock detection and treatment), optimistic concurrency control, timestamp ordering
Extend the concept of transaction in a distributed system
Describe, compare, and implement various distributed commit protocols: one and two-phase
Understand the application of replication in a distributed system, as well as the consistency problems introduced, and describe the corresponding consistency models and their implementation.
Related competences:
CEE2.3,
CEE2.1,
Subcompetences:
Describe specific implementations of consistency models: primary-based protocols (remote-write, local-write) and replicated-write protocols (active replication, quorum-based protocols)
Describe the data-centric strong consistency models: strict, linearizability, sequential
Describe the data-centric relaxed consistency models: usage of synchronization variables
Understand the application of replication in a distributed system, in addition to the consistency problems introduced
Understand the problem of concurrent access to resources by different agents (threads, processors), and the design principles to ensure a correct coordination to avoid deadlocks
Related competences:
CG1,
CEE4.2,
Subcompetences:
Use the formalism of transition systems to describe the semantics of a concurrent system. Employ a modeling tool to analyze properties.
Understand the modeling of a concurrent system, the description and analysis of properties.
Understand and carry out the modeling of a concurrent system using a formal model, and analyze its properties.
Understand the fundamental features to be addressed in a concurrent system: mutual exclusion, critical sections, interference, and the primitives for the design of concurrent programs.
Understand the design of concurrent programs for shared memory architectures
Related competences:
CEE2.3,
CEE4.2,
CTR3,
Subcompetences:
Understand the definition of thread, and the design of concurrent programs that use threads. Understand the basic primitives available in a modern programming language (like Java) to ensure proper coordination between threads.
Understand the problems of interference and blocking in a concurrent system as well as designing concurrent programs free of these problems.
Understand and apply the techniques offered by modern programming languages for the design of concurrent programs: threads, semaphores, critical sections, monitors and synchronization variables.
Understand the design of concurrent programs for message-passing architectures
Related competences:
CEE2.3,
CEE2.1,
CEE4.2,
CTR3,
Subcompetences:
Apply various design patterns of concurrent systems on a message passing architecture: client-server, filter pipeline, worker-supervisor, advertiser-listener.
Understand the paradigm of message-passing programming languages like Erlang. Understand the concept of process in Erlang, and the design of simple / medium-sized programs in Erlang.
Design solutions based on message passing architecture for known problems, such as sorting by quicksort.
Understand and measure the potential parallelism available in a sequential application and the performance achieved with its parallel implementation
Related competences:
CG1,
CEE4.2,
Subcompetences:
Understand the performance of a parallel application and the overhead factors that contribute to its degradation
Understand the concept of task graph and how to measure potential parallelism
Use of instrumentation, visualization and analysis tools to understand parallelism and performance
Decide the most appropriate decomposition strategy to express parallelism in an application
Related competences:
CEE4.2,
CTR3,
CTR6,
CG5,
Subcompetences:
Given a task decomposition, decide task ordering and data sharing constrains that guarantee dependencies
Given a data decomposition, decide communication requirements and the use of either point-to-point or collective operations to implement them
Use of task and data decomposition
Specify, using the appropriate programming paradigm, an efficient parallel version that corresponds to a certain task/data decomposition
Related competences:
CEE4.2,
CTR3,
CTR6,
CG5,
Subcompetences:
Use of the message-passing interface (MPI) for computation/data intensive applications in distributed-memory architectures
Use of CUDA to offload computation and transfer data for accelerator-based architectures
Contents
Transition systems and process algebra
Transition systems and primitive processes. Operations prefix and choice. Concurrent processes and their implementation. Description of the LTS system. Implementation in Java and Erlang.
Understanding parallelism
Introduction to parallel architectures: shared- and distributed-memory architectures and accelerators. Task graph: nodes and edges. Metrics. Speed-up and efficiency. Amdahl law.
Concepts underlying distributed systems
Definition of a distributed system. Potential applications of a distributed system. Examples of distributed systems. Challenges to design and implement a distributed system: heterogeneity, security, lack of global view, concurrency, lack of a single point of control, asynchrony, openness transparency, fault tolerance, scalability.
Safety and liveness properties
Description and examples of safety properties and deployment in LTS. Description of liveness properties, especially the progress property and implementation in LTS.
Concurrent objects, mutual exclusion and synchronization conditions. The deadlock problem.
Problem of destructive interference. Locks and mutual exclusion. Modeling of traffic lights and monitors and of the problem of overlapping. Problem of deadlock, analysis by LTS.
Message passing. Architectures
Message passing. Client / server architecture, other architectures Introduction to.: Pilelines filters, suvervisor / workers, advertiser / listener. Message-passing in Erlang. Design of Erlang in a client / server architecture.
Predicting and analyzing performance
Use of models and tools to understand parallelism and performance (Tareador, Extrae, Paraver and Dimemas).
Distributed-memory programming using MPI
Cluster architecture overview. Process creation, identification and termination. Point-to-point vs. collective operations. Synchronous vs. asynchronous communications.
Programming GPU devices for computation acceleration using CUDA
GPU architecture overview. Decompositions suitable for GPU acceleration. CUDA programming principles. CUDA Parallel Execution Model: host and device.
Distributed algorithms: Time, global state, coordination, and agreement
Time and events ordering in a distributed system. Logical clocks: happened-before relation, Lamport logical clocks (scalar, vector). Algorithms to synchronize physical clocks in a distributed system: Cristian (NTP), Berkeley. Consistent global state in a distributed system. The Chandy and Lamport's mechanism of distributed snapshot. Global predicates for evaluating properties in a distributed system: properties of predicates (stability), occurrence of predicates (possibly, definitely). Coordination of processes in a distributed system for the election of leader: Bully, Ring. Coordination of processes in a distributed system for multicast group communication: basic reliable multicast, scalable reliable multicast, ordered multicast (FIFO, causal, total), atomic multicast. Coordination of processes in a distributed system to ensure consensus: the two army problem, Dolev & Strong's algorithm, the Byzantine generals problem, Paxos
Distributed shared data: Transactions, consistency, and replication
Concurrent execution of transactions: lost update, inconsistent retrievals, serial equivalence, recovery of aborts (dirty read, write premature). Concurrency control mechanisms: two-phase locking (including deadlock detection and treatment), optimistic concurrency control, timestamp ordering. Distributed transaction. Distributed commit protocols: one and two-phase. Concurrency control mechanisms in a distributed system: two-phase locking (including distributed deadlock detection and treatment), optimistic concurrency control, timestamp ordering. Replication and consistency in a distributed system. Data-centered strong consistency models: strict, linearizability, sequential. Data-centric relaxed consistency models: usage of synchronization variables. Client-centric consistency models: eventual, monotonic-read, monotonic-write, read-your-writes, writes follow-reads. Implementations of consistency models: primary-based protocols (remote-write, local-write) and replicated-write protocols (active replication, quorum-based protocols)
Activities
ActivityEvaluation act
Concurrency, parallelism and distribution: fundamental concepts
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents and their subsequent application Objectives:1811 Contents:
First chosen module: Concurrency or Parallelism or Distribution
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents associated to the corresponding module and their subsequent application in the practical sessions.
Theory
10h
Problems
0h
Laboratory
10h
Guided learning
0h
Autonomous learning
30h
Second chosen module: Concurrency or Parallelism or Distribution
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents associated to the corresponding module and their subsequent application in the practical sessions.
Theory
10h
Problems
0h
Laboratory
10h
Guided learning
0h
Autonomous learning
30h
Module I: Concurrency
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents associated to Module I (concurrency) and their subsequent application in the practical sessions. Objectives:8910 Contents:
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents associated to Module II (parallelism) and their subsequent application in the practical sessions. Objectives:111213 Contents:
Class preparation with the help of the support material. Understanding and assimilation of the lesson contents associated to Module III (distribution) and their subsequent application in the practical sessions. Objectives:1234567 Contents:
Assimilation of the concepts of course and the exam Objectives:12345678910111213 Week:
15 (Outside class hours)
Theory
2h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
8h
Second chosen module: Final exam
Assimilation of the concepts of course and the exam Objectives:12345678910111213 Week:
15 (Outside class hours)
Theory
2h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
8h
Teaching methodology
During the course there will be two types of activities:
a) Activities focused on the acquisition of theoretical knowledge.
b) Activities focused on the acquisition of knowledge through experimentation by implementing and evaluating empirically in the laboratory the mechanisms explained at a theoretical level.
The theoretical activities include participatory lecture classes, which explain the basic contents of the course. The practical activities include seminar laboratories where students implement the mechanisms described in the lectures. The seminars require a preparation by reading the statement and supporting documentation, and a further elaboration of the conclusions in a report.
Evaluation methodology
For each module, there will be an exam (EF) and a lab grade (L). The exam will comprise problems on the theory taught. The lab grade will reflect the work done by the students in the practical assignments. The module final grade will be calculated as follows Mi=0.6*EF+ 0.4*L.
The final grade will be calculated as the arithmetic mean of the grades of the two modules taken by the student as follows NF=0.5*M1+ 0.5*M2.
Concurrency: Java at the level of classes and objects.
Parallelism: basic understanding of parallel architectures, including shared- and distributed-memory multiprocessor systems.
Distribution: understanding of the internal structure and operation of an operating system and computer networks.