Skip to content

Commit 79154a3

Browse files
committed
Update documentation
1 parent 59ff3f3 commit 79154a3

5 files changed

Lines changed: 49 additions & 12 deletions

File tree

_sources/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
Pre-trained foundation models have revolutionized speech technology like many other adjacent fields. The combination of their capability and opacity has sparked interest in researchers trying to interpret the models in various ways. While interpretability in fields such as computer vision and natural language processing has made significant progress towards understanding model internals and explaining their decisions, speech technology has lagged behind despite the widespread use of complex, black-box neural models. Recent studies have begun to address this gap, marked by a growing body of literature focused on interpretability in the speech domain. This tutorial provides a structured overview of interpretability techniques, their applications, implications, and limitations when applied to speech models, aiming to help researchers and practitioners better understand, evaluate, debug, and optimize speech models while building trust in their predictions. In hands-on sessions, participants will explore how speech models encode distinct features (e.g., linguistic information) and utilize them in their inference. By the end, attendees will be equipped with the tools and knowledge to start analyzing and interpreting speech models in their own research, potentially inspiring new directions.
66

77
```{note}
8-
We will present our tutorial about _Interpretability Techniques for Speech Models_ on **Sunday, August 17th** at this year's Interspeech conference in Rotterdam. <br> Check out the [programme](#interspeech-programme) below, and browse the materials through the sidebar menu.
8+
We presented our tutorial about _Interpretability Techniques for Speech Models_ on **Sunday, August 17th** at this year's Interspeech conference in Rotterdam. <br> Check out the [programme](#interspeech-programme) below, and browse the materials through the sidebar menu (an overview of all slides and notebooks is [here](https://interpretingdl.github.io/speech-interpretability-tutorial/interspeech2025/intro.html)).
99
```
1010

1111
![tutorial-overview-diagram](images/tutorial-overview.png)

_sources/interspeech2025/intro.md

Lines changed: 14 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,23 @@
11
# Interspeech 2025 tutorial
22

3+
## Slides
4+
5+
- [Grzegorz Chrupała - Introduction](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_GrzegorzChrupa%C5%82a_Introduction.pdf)
6+
- [Martijn Bentum - Representational Analysis techniques](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_MartijnBentum_RepresentationalAnalyses.pdf)
7+
- [Hosein Mohebbi - Context Mixing techniques](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_HoseinMohebbi_ContextMixing.pdf)
8+
- [Gaofei Shen - Feature Attribution techniques](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_GaofeiShen_FeatureAttribution.pdf)
9+
- [Marianne de Heer Kloots - Key takeaways & Future outlook](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_MarianneDeHeerKloots_FutureOutlook.pdf)
10+
- [Willem Zuidema - Discussion](https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_WillemZuidema_Discussion.pdf)
11+
12+
13+
## Notebooks
14+
315
We provide four notebooks, exemplifying the interpretability techniques we cover in the tutorial. The notebooks are designed to work on Google Colab, open them through the links below:
416

5-
## Part I: Representation Understanding
17+
### Part I: Representation Understanding
618
- [Probing exercise: Does Wav2Vec2 encode vocal tract movements?](https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/representational-analyses/probing.ipynb)
719
- [Representation space comparisons: Locating acoustic and articulatory information in Wav2Vec2 with CKA](https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/representational-analyses/representation_space_comparisons.ipynb)
820

9-
## Part II: Feature Importance Scoring
21+
### Part II: Feature Importance Scoring
1022
- [Context Mixing: Quantifying Context-Mixing in Speech Transformers](https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/feature-importance-scoring/context_mixing.ipynb)
1123
- [Feature attribution: Explaining Speech Classification Models with Feature Attribution](https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/feature-importance-scoring/feature_attribution.ipynb)

index.html

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -410,10 +410,13 @@ <h1>Interpretability Techniques for Speech Models<a class="headerlink"
410410
inspiring new directions.</p>
411411
<div class="admonition note">
412412
<p class="admonition-title">Note</p>
413-
<p>We will present our tutorial about <em>Interpretability Techniques for Speech Models</em> on
413+
<p>We presented our tutorial about <em>Interpretability Techniques for Speech Models</em> on
414414
<strong>Sunday, August 17th</strong> at this year’s Interspeech conference in Rotterdam. <br> Check
415415
out the <a class="reference internal" href="#interspeech-programme">programme</a> below, and browse
416-
the materials through the sidebar menu.</p>
416+
the materials through the sidebar menu (an overview of all slides and notebooks is <a
417+
class="reference external"
418+
href="https://interpretingdl.github.io/speech-interpretability-tutorial/interspeech2025/intro.html">here</a>).
419+
</p>
417420
</div>
418421
<p><img alt="tutorial-overview-diagram" src="_images/tutorial-overview.png" /></p>
419422
<section id="programme-at-interspeech-2025">

interspeech2025/intro.html

Lines changed: 28 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -356,8 +356,12 @@ <h2> Contents </h2>
356356
</div>
357357
<nav aria-label="Page">
358358
<ul class="visible nav section-nav flex-column">
359-
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#part-i-representation-understanding">Part I: Representation Understanding</a></li>
360-
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#part-ii-feature-importance-scoring">Part II: Feature Importance Scoring</a></li>
359+
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#slides">Slides</a></li>
360+
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#notebooks">Notebooks</a><ul class="nav section-nav flex-column">
361+
<li class="toc-h3 nav-item toc-entry"><a class="reference internal nav-link" href="#part-i-representation-understanding">Part I: Representation Understanding</a></li>
362+
<li class="toc-h3 nav-item toc-entry"><a class="reference internal nav-link" href="#part-ii-feature-importance-scoring">Part II: Feature Importance Scoring</a></li>
363+
</ul>
364+
</li>
361365
</ul>
362366
</nav>
363367
</div>
@@ -371,21 +375,35 @@ <h2> Contents </h2>
371375

372376
<section class="tex2jax_ignore mathjax_ignore" id="interspeech-2025-tutorial">
373377
<h1>Interspeech 2025 tutorial<a class="headerlink" href="#interspeech-2025-tutorial" title="Link to this heading">#</a></h1>
378+
<section id="slides">
379+
<h2>Slides<a class="headerlink" href="#slides" title="Link to this heading">#</a></h2>
380+
<ul class="simple">
381+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_GrzegorzChrupa%C5%82a_Introduction.pdf">Grzegorz Chrupała - Introduction</a></p></li>
382+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_MartijnBentum_RepresentationalAnalyses.pdf">Martijn Bentum - Representational Analysis techniques</a></p></li>
383+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_HoseinMohebbi_ContextMixing.pdf">Hosein Mohebbi - Context Mixing techniques</a></p></li>
384+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_GaofeiShen_FeatureAttribution.pdf">Gaofei Shen - Feature Attribution techniques</a></p></li>
385+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_MarianneDeHeerKloots_FutureOutlook.pdf">Marianne de Heer Kloots - Key takeaways &amp; Future outlook</a></p></li>
386+
<li><p><a class="reference external" href="https://raw.githubusercontent.com/interpretingdl/speech-interpretability-tutorial/refs/heads/main/book/interspeech2025/Interspeech2025_InterpretabilityTutorial_WillemZuidema_Discussion.pdf">Willem Zuidema - Discussion</a></p></li>
387+
</ul>
388+
</section>
389+
<section id="notebooks">
390+
<h2>Notebooks<a class="headerlink" href="#notebooks" title="Link to this heading">#</a></h2>
374391
<p>We provide four notebooks, exemplifying the interpretability techniques we cover in the tutorial. The notebooks are designed to work on Google Colab, open them through the links below:</p>
375392
<section id="part-i-representation-understanding">
376-
<h2>Part I: Representation Understanding<a class="headerlink" href="#part-i-representation-understanding" title="Link to this heading">#</a></h2>
393+
<h3>Part I: Representation Understanding<a class="headerlink" href="#part-i-representation-understanding" title="Link to this heading">#</a></h3>
377394
<ul class="simple">
378395
<li><p><a class="reference external" href="https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/representational-analyses/probing.ipynb">Probing exercise: Does Wav2Vec2 encode vocal tract movements?</a></p></li>
379396
<li><p><a class="reference external" href="https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/representational-analyses/representation_space_comparisons.ipynb">Representation space comparisons: Locating acoustic and articulatory information in Wav2Vec2 with CKA</a></p></li>
380397
</ul>
381398
</section>
382399
<section id="part-ii-feature-importance-scoring">
383-
<h2>Part II: Feature Importance Scoring<a class="headerlink" href="#part-ii-feature-importance-scoring" title="Link to this heading">#</a></h2>
400+
<h3>Part II: Feature Importance Scoring<a class="headerlink" href="#part-ii-feature-importance-scoring" title="Link to this heading">#</a></h3>
384401
<ul class="simple">
385402
<li><p><a class="reference external" href="https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/feature-importance-scoring/context_mixing.ipynb">Context Mixing: Quantifying Context-Mixing in Speech Transformers</a></p></li>
386403
<li><p><a class="reference external" href="https://colab.research.google.com/github/interpretingdl/speech-interpretability-tutorial/blob/main/book/interspeech2025/feature-importance-scoring/feature_attribution.ipynb">Feature attribution: Explaining Speech Classification Models with Feature Attribution</a></p></li>
387404
</ul>
388405
</section>
406+
</section>
389407
<div class="toctree-wrapper compound">
390408
</div>
391409
</section>
@@ -454,8 +472,12 @@ <h2>Part II: Feature Importance Scoring<a class="headerlink" href="#part-ii-feat
454472
</div>
455473
<nav class="bd-toc-nav page-toc">
456474
<ul class="visible nav section-nav flex-column">
457-
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#part-i-representation-understanding">Part I: Representation Understanding</a></li>
458-
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#part-ii-feature-importance-scoring">Part II: Feature Importance Scoring</a></li>
475+
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#slides">Slides</a></li>
476+
<li class="toc-h2 nav-item toc-entry"><a class="reference internal nav-link" href="#notebooks">Notebooks</a><ul class="nav section-nav flex-column">
477+
<li class="toc-h3 nav-item toc-entry"><a class="reference internal nav-link" href="#part-i-representation-understanding">Part I: Representation Understanding</a></li>
478+
<li class="toc-h3 nav-item toc-entry"><a class="reference internal nav-link" href="#part-ii-feature-importance-scoring">Part II: Feature Importance Scoring</a></li>
479+
</ul>
480+
</li>
459481
</ul>
460482
</nav></div>
461483

searchindex.js

Lines changed: 1 addition & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)