Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

Conversation

NikolaiKyhne
Copy link

@NikolaiKyhne NikolaiKyhne commented Sep 9, 2024

Added Multi-Window Multi-Head attention (MWMHA) module for Transformer ASR (https://openreview.net/forum?id=Q53QLftNkA).

In general, this contribution adds:

  • MWMHA implementation
  • Updated Transformer.py with MWMHA option
  • Updated TransformerASR.py with MWMHA option
  • Added small and medium MWMHA LibriSpeech recipe
  • Added large MWMHA CommonVoice english recipe
  • Updated LibriSpeech and CommonVoice ASR README with MWMHA results and description

@Adel-Moumen Adel-Moumen marked this pull request as draft September 10, 2024 13:29
@Adel-Moumen
Copy link
Collaborator

Hey guys!

Hope you are doing great --- this is a very nice PR!

I just turned this PR draft for now, please turn it public when you think it will be ready to be reviewed. You can ping me as well so that I can have a closer look as soon as possible :)

Thanks for your contribution :)

Best,
Adel

@NikolaiKyhne NikolaiKyhne marked this pull request as ready for review September 11, 2024 17:46
@NikolaiKyhne
Copy link
Author

NikolaiKyhne commented Sep 11, 2024

Hey @Adel-Moumen!

Thanks for your comment, we have now finished the draft and turned it ready for review :)

Best,
Nikolai

## Transformer
| Language | CV version | hyperparams file | LM | Val. CER | Val. WER | Test CER | Test WER | Hugging Face link | Model link | GPUs |
| ------------- |:-------------:|:---------------------------:| -----:| -----:| -----:| -----:| -----:|:-----------:| :-----------:| :-----------:|
| English | 16.1 | mwmha_transformer_large.yaml | No | 4.72 | 10.97 | 6.68 | 13.69 | - | [model](https://1drv.ms/f/c/039f8ffe91e06416/Et7KEbSlWNdJhkjLIi7_vGQBMVhGwRRBzCSljh6aA4sJSw?e=dXeuiY) | 1xL40 48GB |
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why is the Val WER so high? I think you swapped CER and WER right ?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

No that's right I just double checked, it is the same for Conformer English on CV 16.1 :)

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@Adel-Moumen Val WER for MWMHA (10.97) follows the same trend and is quite close to that of the Conformer model (10.48) and is reported correctly, CER and WER are not swapped.

@SarthakYadav
Copy link
Contributor

Hi @Adel-Moumen @TParcollet

We've been waiting for a review for some time now. Any chance you can take a look at it soon? :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants

Morty Proxy This is a proxified and sanitized view of the page, visit original site.