Jay Alammar 12/17/2020

Interfaces for Explaining Transformer Language Models

Read Original

This article presents interactive visualizations and explorables for explaining transformer-based language models, such as GPT-2. It covers input saliency methods to score token importance and neuron activation analysis to understand how model components generate outputs. It is the first in a series on model interpretability and is accompanied by an open-source library (Ecco) for creating similar interfaces in Jupyter notebooks.

Interfaces for Explaining Transformer Language Models

Comments

No comments yet

Be the first to share your thoughts!

Browser Extension

Get instant access to AllDevBlogs from your browser