Skip to main content

Introduction to Human Language Technology

Credits
5
Types
Compulsory
Requirements
This subject has not requirements , but it has got previous capacities
Department
CS
Web
www.cs.upc.edu/~turmo/ihlt/plan32js/IHLT.html
The goal of this course is to provide the fundamentals of Natural Language Processing (NLP) to the student. Concretely, the course is an introduction to the most relevant drawbacks involved in NLP, the most relevant techniques and resources used to tackle with them, and the theories they are based on. In addition, brief descriptions of the most relevant NLP applications are included. The course will focus on knowledge-based and empirical-based approaches to NLP (both statistical and machine learning).

IHLT provides the basic NLP knowledge in order to course AHLT and HLE. While AHLT goes in depth in the NLP statistical techniques, HLE reviews the state of the art on real applications in which NLP technology is involved.

Teachers

Person in charge

Others

Weekly hours

Theory
2
Problems
0
Laboratory
1
Guided learning
0
Autonomous learning
5.93

Competences

Generic

  • CG1 - Capability to plan, design and implement products, processes, services and facilities in all areas of Artificial Intelligence.
  • CG3 - Capacity for modeling, calculation, simulation, development and implementation in technology and company engineering centers, particularly in research, development and innovation in all areas related to Artificial Intelligence.
  • Academic

  • CEA5 - Capability to understand the basic operation principles of Natural Language Processing main techniques, and to know how to use in the environment of an intelligent system or service.
  • Professional

  • CEP4 - Capability to design, write and report about computer science projects in the specific area of ??Artificial Intelligence.
  • CEP6 - Capability to assimilate and integrate the changing economic, social and technological environment to the objectives and procedures of informatic work in intelligent systems.
  • CEP7 - Capability to respect the legal rules and deontology in professional practice.
  • Teamwork

  • CT3 - Ability to work as a member of an interdisciplinary team, as a normal member or performing direction tasks, in order to develop projects with pragmatism and sense of responsibility, making commitments taking into account the available resources.
  • Information literacy

  • CT4 - Capacity for managing the acquisition, the structuring, analysis and visualization of data and information in the field of specialisation, and for critically assessing the results of this management.
  • Reasoning

  • CT6 - Capability to evaluate and analyze on a reasoned and critical way about situations, projects, proposals, reports and scientific-technical surveys. Capability to argue the reasons that explain or justify such situations, proposals, etc..
  • Objectives

    1. Understand the fundamental concepts of Natural Language Processing, most well-known techniques and theories as well as most relevant existing resources.
      Related competences: CEA5, CG1, CG3, CEP6, CT4, CT6,
    2. Understand most relevant applications of NLP and the theories, tecniques and resources they use.
      Related competences: CEA5, CG1, CG3, CEP6, CT4, CT6,
    3. Design and development of programs to solve specific problems in the NLP context, involving the selection of most appropiate techniques and resources as well as the use of existing resources. There would be one larger programs to be developed in groups of two students.
      Related competences: CEA5, CG1, CG3, CEP4, CEP6, CEP7, CT3, CT4, CT6,
    4. Reason (ocassionally, in group) about several problems in the NLP context that imply considering different techniques and resources.
      Related competences: CEA5, CG1, CG3, CEP7, CT3, CT4, CT6,

    Contents

    1. Document Structure and Language
      Text selection, Tokenization, Sentence splitting, Language Identifiers
    2. Words
      Morphology, Finite States Automata, Finite States Transducers.
      PoS tagging, Hidden Markov Models.
      Lexical semantics, Semantic resources.
      Word Sense Diambiguation.
    3. Word sequences
      Recognition and classification of word sequences with meaning.
      BIO discriminative models. Conditional Random Fields (CRF).
      Named Entity Recognition and Classification (NERC).
      Noun-phrase Chunking.
    4. Sentences
      Syntactic grammars, typology. Context free grammars. Probabilistic context free grammars. Chomsky normal form grammars.

      Syntactic parsers, properties and strategies. CKY and probabilistic CKY parsers.
    5. Sentence sequences
      Coreference resolution. Mention detection. Types of techniques for the generation of coreferents chains. Mention-pair model. Entity-mention model. Rankers model.

    Activities

    Activity Evaluation act


    Introduction


    Objectives: 1 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Document structure and language


    Objectives: 1 3
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Morphological analysis

    Finite States Automata. Finite States Transducers.
    Objectives: 1 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    PoS tagging

    Hidden Markov Models
    Objectives: 1 4 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Lexical semantics, Semantic resources.


    Objectives: 1 4 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Word Sense Diambiguation.


    Objectives: 1 4 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Recognition and classification of word sequences with meaning.

    BIO discriminative models. Conditional Random Fields (CRF). Named Entity Recognition and Classification (NERC). Noun-phrase Chunking.
    Objectives: 4 3 1
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Syntactic parsing: Syntactic grammars

    Typology. Context free grammars. Probabilistic context free grammars. Chomsky normal form grammars.
    Objectives: 1 4 2
    Theory
    2h
    Problems
    0h
    Laboratory
    2h
    Guided learning
    0h
    Autonomous learning
    0h

    Syntactic parsing: parsers

    Syntactic parsers, properties and strategies. CKY and probabilistic CKY parsers.
    Objectives: 1 4 2
    Theory
    4h
    Problems
    0h
    Laboratory
    2h
    Guided learning
    0h
    Autonomous learning
    0h

    Coreference resolution


    Objectives: 1 2
    Theory
    2h
    Problems
    0h
    Laboratory
    1h
    Guided learning
    0h
    Autonomous learning
    0h

    Project tutoring


    Objectives: 4 2
    Contents:
    Theory
    0h
    Problems
    0h
    Laboratory
    2h
    Guided learning
    0h
    Autonomous learning
    0h

    Project presentation



    Theory
    4h
    Problems
    0h
    Laboratory
    0h
    Guided learning
    0h
    Autonomous learning
    40h

    Final exam



    Week: 15 (Outside class hours)
    Theory
    0h
    Problems
    0h
    Laboratory
    0h
    Guided learning
    0h
    Autonomous learning
    0h

    Teaching methodology

    There are two types of sessions: theory/exercise and laboratory.

    In each theory/exercise session we will introduce new concepts together with the challenges they present and the approaches to face them. In addition, we will solve some exercises to fix those concepts, techniques and algorithms introduced in the session.

    In the laboratory sessions small practices will be developed using the appropriate NLP tools to practice and reinforce the knowledge learned in the theory classes.

    Evaluation methodology

    There will be a unique exam at the end of the course, one project and one deliverable for each lab session. The exam will include all the course contents.
    The mark of the project and deliverables will be computed by considering the documents presented by the students.
    The final mark of the course will be calculated as follows:
    Course mark = final exam mark* 0.5 + lab mark * 0.5

    Bibliography

    Basic

    Web links

    Previous capacities

    Those acquired in the course of Artificial Intelligence (AI) (degree in Computer Engineering)