Credits
6
Types
Compulsory
Requirements
This subject has not requirements
, but it has got previous capacities
Department
CS
Web
https://sites.google.com/upc.edu/gia-iaa/
Teachers
Person in charge
- Sergio Álvarez Napagao ( salvarez@cs.upc.edu )
Others
- Jordi Luque Serrano ( jordi.luque.serrano@upc.edu )
Weekly hours
Theory
2
Problems
0
Laboratory
2
Guided learning
0
Autonomous learning
6
Competences
Transversals
Basic
Especifics
Generic
Objectives
-
Learn the main methods of machine learning, and how to use them appropriately.
Related competences: CG1, CG2, CG3, CG4, CG8, CT2, CT5, CB3, CE03, CE04, CE09, CE15, CE20, -
Interact in a critical and prudent manner with data and machine learning models
Related competences: CG1, CG4, CG7, CG8, CG9, CT6, CT8, CE04,
Subcompetences- Keep a critic and skeptic view of model behavior
- Identify biases in data
-
Recognize in an easy manner the characteristics of a problem from the perspective of machine learning
Related competences: CG1, CG2, CG4, CG6, CG9, CT5, CE03, CE04, CE09, CE15,
Subcompetences- Identify analysis of relevance to be conducted on a data set
- Propose the most appropriate learning types for a problem
Contents
-
Intro to machine learning.
Basic tipes of learning. What can they be used for, purposes and main limitations. Includes a set of warnings and sanity checks to keep in mind while working with machine learning. -
Experimental design in machine learning
Using data for learning. How to design, execute and evaluate experiments conducted with machine learning techniques. -
Data preprocessing
Distributions, normalization and standardization of data. How and why to prepare data to be processed by machine learning algorithms. -
Applied Regression
Practical cases of regression -
Dimensionality reduction
Review of the main methods to reduce the dimensionality of data: PCA, UMAP, T-SNE, ... -
Classification: Basic concepts and review of basic methods
Distance measurements are studied and related to the concept of similitude, which then allows us to build and compare a large number of methods. Revision of the K-Nearest Neighbor as a simple framework and extension to other methods. -
Classification methods based on other criteria
Support Vector Machines, Neural Networks (classic architectures) and Decision Trees. -
Muilticlassification
The main methods of combining "weak" learning methods are studied in order to obtain more robust models: Boosting, Bagging, GAMs, EBMs, Sets -
Explainability
Relevance, use and methods of explicability. Several methods are studied in order to interpret and explain the operation and result of machine learning algorithms, a basic need for the implementation and acceptance of these methods. The foundations for Explainable AI (Explainable Artificial Intelligence) are being laid. -
Clustering
The bases of the classical methods of obtaining significant data sets in the absence of class information and / or prior structures are reviewed. K-means, Hierarchical Clustering, Spectral Clustering, DBSCAN. -
Genetic algorisms
Introduction to genetic algorithms, as a first vision of bio-inspired learning methods. The conceptual and mathematical bases of the main mutation and crossover operators and their representational variants are reviewed. -
Machine learning in graphs
The structure of the graph is widespread in various environments and has donated to a whole discipline, the Science of the Xarxes, on it is work on the structural properties of the graphs to derive properties and conclusions about the phenomenon or field that is studied. This type of learning is especially important in internet applications, near or recovery applications or knowledge detection. Detection of communities, prediction of accidents, etc.
Activities
Activity Evaluation act
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h
Teaching methodology
Interactive classes of theoretical content. Relatively autonomous laboratory sessions of practical contingency.Evaluation methodology
The course consists of one partial exam (P) and a final exam (F). The laboratory will be evaluated continuously (LC) and through a final delivery (LF).Final score = (0.2*P) + (0.4*F) + (0.1*LC) + (0.3*LF)
Reassessment: Only those who have failed the final exam may take the reassessment. The maximum grade that can be obtained in the reassessment is 7.
Bibliography
Basic
-
Pattern recognition and machine learning
- Bishop, Christopher M,
Springer,
cop. 2006.
ISBN: 0387310738
https://discovery.upc.edu/discovery/fulldisplay?docid=alma991003157379706711&context=L&vid=34CSUC_UPC:VU1&lang=ca
Previous capacities
Understand the computing flow within a software system.Understand the basic concepts behind inference, deduction and evidence based reasoning.
Being familiarized with data distribution, basic data preprocessing, i how numerical variables can represent information.