Introduction to Human Language Technology

You are here

Credits
5
Types
Compulsory
Requirements
This subject has not requirements, but it has got previous capacities
Department
CS
The goal of this course is to provide the fundamentals of Natural Language Processing (NLP) to the student. Concretely, the course is an introduction to the most relevant drawbacks involved in NLP, the most relevant techniques and resources used to tackle with them, and the theories they are based on. In addition, brief descriptions of the most relevant NLP applications are included. The course will focus on knowledge-based and empirical-based approaches to NLP (both statistical and machine learning).

IHLT provides the basic NLP knowledge in order to course AHLT and HLE. While AHLT goes in depth in the NLP statistical techniques, HLE reviews the state of the art on real applications in which NLP technology is involved.

Teachers

Person in charge

  • Jordi Turmo Borrás ( )

Others

  • Gerard Escudero Bakx ( )
  • Salvador Medina Herrera ( )

Weekly hours

Theory
2
Problems
0
Laboratory
1
Guided learning
0.21
Autonomous learning
8.93

Competences

Generic Technical Competences

Generic

  • CG1 - Capability to plan, design and implement products, processes, services and facilities in all areas of Artificial Intelligence.
  • CG3 - Capacity for modeling, calculation, simulation, development and implementation in technology and company engineering centers, particularly in research, development and innovation in all areas related to Artificial Intelligence.

Technical Competences of each Specialization

Academic

  • CEA5 - Capability to understand the basic operation principles of Natural Language Processing main techniques, and to know how to use in the environment of an intelligent system or service.

Professional

  • CEP4 - Capability to design, write and report about computer science projects in the specific area of ??Artificial Intelligence.
  • CEP6 - Capability to assimilate and integrate the changing economic, social and technological environment to the objectives and procedures of informatic work in intelligent systems.
  • CEP7 - Capability to respect the legal rules and deontology in professional practice.

Transversal Competences

Teamwork

  • CT3 - Ability to work as a member of an interdisciplinary team, as a normal member or performing direction tasks, in order to develop projects with pragmatism and sense of responsibility, making commitments taking into account the available resources.

Information literacy

  • CT4 - Capacity for managing the acquisition, the structuring, analysis and visualization of data and information in the field of specialisation, and for critically assessing the results of this management.

Reasoning

  • CT6 - Capability to evaluate and analyze on a reasoned and critical way about situations, projects, proposals, reports and scientific-technical surveys. Capability to argue the reasons that explain or justify such situations, proposals, etc..

Objectives

  1. Understand the fundamental concepts of Natural Language Processing, most well-known techniques and theories as well as most relevant existing resources.
    Related competences: CEA5, CG1, CG3, CEP6, CT4, CT6,
  2. Understand most relevant applications of NLP and the theories, tecniques and resources they use.
    Related competences: CEA5, CG1, CG3, CEP6, CT4, CT6,
  3. Design and development of programs to solve specific problems in the NLP context, involving the selection of most appropiate techniques and resources as well as the use of existing resources. There would be one larger programs to be developed in groups of two students.
    Related competences: CEA5, CG1, CG3, CEP4, CEP6, CEP7, CT3, CT4, CT6,
  4. Reason (ocassionally, in group) about several problems in the NLP context that imply considering different techniques and resources.
    Related competences: CEA5, CG1, CG3, CEP7, CT3, CT4, CT6,

Contents

  1. Document Structure and Language
    Text selection, Tokenization, Sentence splitting, Language Identifiers
  2. Words
    Morphology, Finite States Automata, Finite States Transducers.
    PoS tagging, Hidden Markov Models.
    Lexical semantics, Semantic resources.
    Word Sense Diambiguation.
  3. Word sequences
    Recognition and classification of word sequences with meaning.
    BIO discriminative models. Conditional Random Fields (CRF).
    Named Entity Recognition and Classification (NERC).
    Noun-phrase Chunking.
  4. Sentences
    Syntactic grammars, typology. Context free grammars. Probabilistic context free grammars. Chomsky normal form grammars.

    Syntactic parsers, properties and strategies. CKY and probabilistic CKY parsers.
  5. Sentence sequences
    Coreference resolution. Mention detection. Types of techniques for the generation of coreferents chains. Mention-pair model. Entity-mention model. Rankers model.

Activities

Activity Evaluation act


Introduction


Objectives: 1 2
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
3h

Document structure and language


Objectives: 1 3
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
8h

Morphological analysis

Finite States Automata. Finite States Transducers.
Objectives: 1 2
Theory
3h
Problems
0h
Laboratory
3h
Guided learning
0h
Autonomous learning
16h

PoS tagging

Hidden Markov Models
Objectives: 1 4 2
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
8h

Lexical semantics, Semantic resources.


Objectives: 1 4 2
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
16h

Word Sense Diambiguation.


Objectives: 1 4 2
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
16h

Recognition and classification of word sequences with meaning.

BIO discriminative models. Conditional Random Fields (CRF). Named Entity Recognition and Classification (NERC). Noun-phrase Chunking.
Objectives: 4 3 1
Theory
3h
Problems
0h
Laboratory
3h
Guided learning
0h
Autonomous learning
16h

Syntactic parsing: Syntactic grammars

Typology. Context free grammars. Probabilistic context free grammars. Chomsky normal form grammars.
Objectives: 1 4 2
Theory
3h
Problems
0h
Laboratory
3h
Guided learning
0h
Autonomous learning
16h

Syntactic parsing: parsers

Syntactic parsers, properties and strategies. CKY and probabilistic CKY parsers.
Objectives: 1 4 2
Theory
3h
Problems
0h
Laboratory
3h
Guided learning
0h
Autonomous learning
16h

Coreference resolution


Objectives: 1 2
Theory
1.5h
Problems
0h
Laboratory
1.5h
Guided learning
0h
Autonomous learning
8h

Project presentation


Objectives: 4 3
Week: 16
Type: assigment
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0.5h
Autonomous learning
2h

Final Exam


Objectives: 1 4 2
Week: 15 (Outside class hours)
Type: final exam
Theory
0h
Problems
0h
Laboratory
0h
Guided learning
0h
Autonomous learning
0h

Teaching methodology

There are two types of sessions: theory/exercise and laboratory.

In each theory/exercise session we will introduce new concepts together with the challenges they present and the approaches to face them. In addition, we will solve some exercises to fix those concepts, techniques and algorithms introduced in the session.

In the laboratory sessions small practices will be developed using the appropriate NLP tools to practice and reinforce the knowledge learned in the theory classes.

Evaluation methodology

There will be a unique exam at the end of the course, one project adn a deliverrable for each lab session. The exam will include all the course contents.
The mark of the project and deliverables will be computed by considering the documents presented by the students.
The final mark of the course will be calculated as follows:
Course mark = final exam mark* 0.5 + lab mark * 0.5

Bibliography

Basic:

Web links

Previous capacities

Those acquired in the course of Artificial Intelligence (AI) (degree in Computer Engineering)