Skriv ut Lukk vindu


 

Vår 2023

INF-8605 Interpretability in Deep Learning - 5 stp


The course is administrated by

Institutt for informatikk

Type of course

The course can be taken as a singular course. Registration is open for UiT students and members of NORA Research School and Digital Life Norway Research School. It will be conducted as a concentrated course in the style of summer- and winter school courses conducted under NORA/DLN, etc.

NOTE: First lecture will be in the beginning of May and will be given digitally.


Course overlap

INF-3605 Interpretability in Deep Learning 4 stp

Course contents

This course will consider different topics of importance regarding interpretable deep learning, equipping the students with knowledge of approaches that can be used to explain deep learning, and deep learning approaches that are more explainable than others. In addition, the students will receive practical skills of applying selected approaches for explaining deep learning, which will equip the students with practical skills of adapting to the rapid pace of technology development in the field of explainable artificial intelligence/interpretable deep learning.

Introductory concepts (3 hours)

Model-agnostic approaches (6 hours)

Neural networks and explainability (6 hours)

Self-reading

Extensive lab work, self-exercises and groups are planned for competence development is also included.

Relevance of course in program of study: Artificial intelligence(AI) and deep learning approaches are often considered as black boxes, i.e. as a type of algorithms that accomplish learning tasks but cannot explain how. However, as AI/deep learning is increasingly absorbed as adopted for accomplishing cognitive tasks for human beings, it is becoming important that the deep learning models are understandable by humans, such that artificial and human intelligence can co-exist and collaborate. In critical tasks such as deriving, from given data, a correct medical diagnosis and prognosis, collaboration between artificial and human intelligence in imperative so that the suggestions or decision from artificial intelligence are both more accurate and more trustworthy.


Admission requirements

PhD students or holders of a Norwegian master´s degree of five years (300 ECTS) or 3 (180 ECTS) + 2 years (120 ECTS) or equivalent may be admitted. PhD students must upload a document from their university stating that there are registered PhD students. This group of applicants does not have to prove English proficiency and are exempt from semester fee. Holders of a Master´s degree must upload a Master´s Diploma with Diploma Supplement / English PhD students at UiT The Arctic University of Norway register for the course through StudentWeb. External applicants apply for admission through SøknadsWeb. All external applicants have to attach a confirmation of their status as a PhD student from their home institution. Students who hold a Master of Science degree, but are not yet enrolled as a PhD-student have to attach a copy of their master's degree diploma. These students are also required to pay the semester fee.

Recommended prerequisites: Programming skills in python and / or INF-1400. Hands on knowledge of python programming for deep learning.

Application code: 9303

Application deadline: March 15th, 2023.

The course is limited to 60 places. Qualified applicants are ranked on the basis of a lottery if there are more applicants than available places.


Objective of the course

A general educational aim of the course will be to equip students with knowledge and skills regarding interpretable artificial intelligence/deep learning, for considering explainable approaches for solving a neural learning problem as well as for developing an explanation for an existing knowledge model developed by a black box approach. This will enable the students to understand, work with, and solve deep learning tasks with a balance of explainability and accuracy, as needed.

An brief introduction to explainable / interpretable deep learning is being offered in this course. This will fill the knowledge gap for those who want to learn more about deep learning and develop trustworthy, reliable deep learning models. Recognizing the significance of interpretable models for computationally-intensive deep learning architectures, as well as the analysis and comprehension of complex biological applications, and the need for cross-disciplinary collaborations in future biotech, medicine, and AI.

PhD student undertaking this course knowledge and learning will be evaluated in an oral exam during their project presentation.

Knowledge - The student

Skills - The student can

General competence - The student has developed


Language of instruction

The language of instruction is English, and all the syllabus material is in English. Project presentation should be given in English and Q&A must be answered in English.

Teaching methods

Lectures: 15 hours

Self-study session: 30 hours

Project work: spread over 8 weeks - net time 50 hours

Group / Self-work session: 20 hours

Hands-on session: 3 hours

Project consultation session: 4 hours

Oral Presentations/ Presentation preparation: 3 hours

Net effort (~125 hours)

Note! First lecture will be in the beginning of May and will be given digitally.