Unsupervised and Reinforcement Learning

You are here

Credits
6
Types
Compulsory
Requirements
This subject has not requirements, but it has got previous capacities
Department
CS
This course covers two important fields of machine learning: non-supervised learning and reinforcement learning. Non-supervised learning is a type of machine learning where the algorithm learns patterns and structures from unlabeled data, whereas reinforcement learning is a type of machine learning where the algorithm learns from feedback given in the form of rewards or punishments.

The course will start with an introduction to the fundamental concepts and algorithms of non-supervised deep learning, such as autoencoders, adversarial networks or denosiing diffuson. The course will then move on to reinforcement learning, covering concepts such as Markov Decision Processes, Q-Learning, and Policy Gradient methods. The course will also explore the latest research in these fields, including deep reinforcement learning and unsupervised deep learning.

By the end of the course, students will have a strong foundation in non-supervised and reinforcement learning, and will be able to apply these techniques to real-world problems.

Teachers

Person in charge

  • Javier Béjar Alonso ( )
  • Mario Martín Muñoz ( )

Weekly hours

Theory
2
Problems
0
Laboratory
2
Guided learning
0
Autonomous learning
6

Competences

Transversal Competences

Transversals

  • CT6 [Avaluable] - Autonomous Learning. Detect deficiencies in one's own knowledge and overcome them through critical reflection and the choice of the best action to extend this knowledge.

Basic

  • CB5 - That the students have developed those learning skills necessary to undertake later studies with a high degree of autonomy

Technical Competences

Especifics

  • CE18 - To acquire and develop computational learning techniques and to design and implement applications and systems that use them, including those dedicated to the automatic extraction of information and knowledge from large volumes of data.

Generic Technical Competences

Generic

  • CG2 - To use the fundamental knowledge and solid work methodologies acquired during the studies to adapt to the new technological scenarios of the future.
  • CG4 - Reasoning, analyzing reality and designing algorithms and formulations that model it. To identify problems and construct valid algorithmic or mathematical solutions, eventually new, integrating the necessary multidisciplinary knowledge, evaluating different alternatives with a critical spirit, justifying the decisions taken, interpreting and synthesizing the results in the context of the application domain and establishing methodological generalizations based on specific applications.

Objectives

  1. To distinguish the kind of problems can be modeled as a reinforcement learning problem and identify the techniques that can be applied to solve them.
    Related competences: CG2, CT6, CE18,
  2. To understand the need, fundamentals, and particularities of behavior learning and the differences it has from supervised and unsupervised machine learning.
    Related competences: CG2, CE18,
  3. To understand the most important algorithms and state of the art in the area of learning by reinforcement
    Related competences: CG4, CE18,
  4. To know how to computationally formalize a real world problem as learning by reinforcement and know how to implement in the most current environments the learning algorithms that solve them
    Related competences: CG2, CG4, CT6, CE18,
  5. Know the problems that can be modeled with deep unsupervised algorithms
    Related competences: CG2, CT6, CE18,
  6. Understand the particularities of deep unsupervised algorithms
    Related competences: CG4, CT6, CE18,
  7. Know the most important algorithms and the state of the art of deep unsupervised learning
    Related competences: CG2, CT6, CB5, CE18,
  8. Knowing how to implement and apply deep learning algorithms to a problem using the most current environment
    Related competences: CG2, CT6, CB5, CE18,

Contents

  1. Introduction: Behavior Learning in Agents and description of main elements in Reinforcement Learning
    Intuition, motivation and definition of the reinforcement learning (RL) framework. Key elements in RL.
  2. Finding optimal policies using Dynamic Programming
    How to learn the optimal policy with full knowledge of the world model: algebraic solution, policy iteration and value iteration.
  3. Introduction to Model-Free approaches.
    Basic algorithms for reinforcement learning: Monte-Carlo, Q-learning, Sarsa, TD(lambda). The need for Exploration. Differences between On-policy and Off-policy methods.
  4. Function approximation in Reinforcement Learning
    Need for function approximation and Incremental methods in RL. The Gradient Descent approach. RL with Linear function approximation. The deadly triad for function approximation in RL. Batch methods and Neural Networks for function Approximation.
  5. Deep Reinforcement Learning (DRL)
    Introducing Deep Learning in RL. Dealing with the deadly triad with the DQN algorithm. Application to the Atari games case. Evolutions of the DQN algorithm: Double DQN, Prioritized Experience Replay, multi-step learning and Distributional value functions. Rainbow: the state-of-the-art algorithm in discrete action space.
  6. Policy gradient methods
    What to do in continuous action spaces. How probabilistic policies allow to apply the gradient method directly in the policy network. The REINFORCE algorithm. The Actor-Critic algorithms. State-of-the-art algorithms in continuous action spaces: DDPG, TD3 and SAC.
  7. Advanced Topics: How to deal with sparse rewards
    The problem of the sparse reward. Introduction to advanced exploration techniques: curiosity and empowerment in RL. Introduction to curriculum learning to easy the learning of the goal. Hierarchical RL to learn complex tasks. The learning of Universal Value Functions and Hindsight Experience Replay (HER).
  8. Reinforcement Learning in the multi-agent framework
    Learning of behaviors in environment where several agents act. Learning of cooperative behaviors, Learning of competitive behaviors, and mixed cases. State-of-the art algorithms. The special case of games: The AlfaGo case and the extension to Alfa-Zero.
  9. Introduction: Deep unsupervised learning
    Introduction to the need for deep unsupervised learning and its applications
  10. Autoregressive models
    Introduction to learning probability distributions defined as autoregressive distributions and main models
  11. Normalizing flows
    Introduction to normalized flows for learning probability distributions
  12. Latent variables models
    Introduction to models based on latent variables and variational autoencoders
  13. Generative Adversarial Networks
    Introduction to generative adversarial networks, conditional and unconditional generation, attribute disentanglement
  14. Denoising Diffusion netwoks
    Introduction to models based on noise diffusion, denoising networks, conditioning, multimodal generation
  15. Self supervised learning
    Introduction to self-supervised learning for feature-generating networks and embeddings, contrastive and non contrastive methods, masking

Activities

Activity Evaluation act


Introduction: Behavior Learning in Agents and description of main elements in Reinforcement Learning



Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Finding optimal policies using Dynamic Programming

How to learn the optimal policy with full knowledge of the world model: algebraic solution, policy iteration and value iteration.

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Introduction to Model-Free approaches. Monte-Carlo, Q-learning, Sarsa, TD(lambda)

Development of the corresponding topic

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Function approximation in RL



Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Deep Reinforcement Learning (DRL)



Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Policy gradient methods

What to do in continuous action spaces. How probabilistic policies allow to apply the gradient method directly in the policy network. The REINFORCE algorithm. The Actor-Critic algorithms. State-of-the-art algorithms in continuous action spaces: DDPG, TD3 and SAC.

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Advanced Topics: How to deal with sparse rewards

The problem of the sparse reward. Introduction to advanced exploration techniques: curiosity and empowerment in RL. Introduction to curriculum learning to easy the learning of the goal. Hierarchical RL to learn complex tasks. The learning of Universal Value Functions and Hindsight Experience Replay (HER).

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Reinforcement Learning in the multi-agent framework

Learning of behaviors in environment where several agents act. Learning of cooperative behaviors, Learning of competitive behaviors, and mixed cases. State-of-the art algorithms. The special case of games: The AlfaGo case and the extension to Alfa-Zero.

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
9h

Control of the reinforcement learning part


Objectives: 3 4 2 1
Week: 8 (Outside class hours)
Type: theory exam
Theory
2h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h

Introduction: Deep unsupervised learning

Introduction to the need for deep unsupervised learning and its applications

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Autoregressive models

Introduction to learning probability distributions defined as autoregressive distributions and main models

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Normalizing flows

Introduction to normalized flows for learning probability distributions

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Latent variables models

Introduction to models based on latent variables and variational autoencoders

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Generative Adversarial Networks

Introduction to generative adversarial networks, conditional and unconditional generation, attribute disentanglement

Theory
2h
Problems
0h
Laboratory
2h
Guided learning
0h
Autonomous learning
6h

Denoising Diffusion netwoks and Self supervised learning

Introduction to models based on noise diffusion, denoising networks, conditioning, multimodal generation

Theory
2h
Problems
0h
Laboratory
4h
Guided learning
0h
Autonomous learning
9h

Unsupervised learning syllabus control


Objectives: 5 6 7 8
Week: 15 (Outside class hours)
Type: theory exam
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h

Teaching methodology

The classes are divided into theory, problem and laboratory sessions.

In the theory sessions, knowledge of the subject will be developed, interspersed with the presentation of new theoretical material with examples and interaction with the students in order to discuss the concepts.

In the laboratory classes, small practices will be developed using tools and using specific libraries that will allow you to practice and reinforce the knowledge of the theory classes.

Evaluation methodology

The subject will include the following assessment acts:

- Reports of the laboratory activities, which must be delivered within the deadline indicated for each session (roughly, 2 weeks). Based on a weighted average of the grades of these reports, a laboratory grade will be calculated, L.

- A first partial exam, taken towards the middle of the course, of the material seen until then. Let P1 be the grade obtained in this exam.

- On the designated day within the exam period, a second partial exam of the subject not covered by the first partial. Let P2 be the grade obtained in this exam.

The three grades L, P1, P2 are between 0 and 10.

The final grade of the subject will be: 0.4*L +0.3*P1 + 0.3*P2

Bibliography

Basic:

Complementary:

Previous capacities

Basic knowledge of Deep Learning and Machine Learning.