Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -25,20 +25,20 @@ We host several models, which are specifically tailored to the processing of Fle
|
|
| 25 |
|
| 26 |
### Automatic Speech Recognition (ASR)
|
| 27 |
|
| 28 |
-
-- **NeLF_S2T_Pytorch** (Recommended): The third version of our Automatic Speech Recognition and Subtitle Generation model. It is a fine-tuned version of ASR_subtitles_v2 without Kaldi-dependency (pure Pytorch), and refined training data leveraging contextualisation techniques for pseudo-labeling.
|
| 29 |
|
| 30 |
-
-- **ASR_subtitles_v2
|
| 31 |
It can generate both an exact verbatim transcription with annotation tags as well as a fully formatted and cleaned up subtitle transcription.
|
| 32 |
|
| 33 |
-
-- **ASR_subtitles_v2_small
|
| 34 |
|
| 35 |
-
-- **ASR_subtitles_v1
|
| 36 |
|
| 37 |
-
-- **ASR_verbatim_v1
|
| 38 |
|
| 39 |
-- **Whisper**: A finetuned Whisper Large model on Flemish data can be found [here](https://huggingface.co/kul-speech-lab/whisper_large_CGN). Usage instructions can be found in Whisper documentation.
|
| 40 |
|
| 41 |
-
**USAGE**: To use our ASR models and transcribe speech yourself, use [our codebase](https://github.com/nelfproject/NeLF_Transcription_ASR).
|
| 42 |
|
| 43 |
### Speaker Diarization and Identification
|
| 44 |
|
|
|
|
| 25 |
|
| 26 |
### Automatic Speech Recognition (ASR)
|
| 27 |
|
| 28 |
+
-- [**NeLF_S2T_Pytorch**](https://huggingface.co/nelfproject/NeLF_S2T_Pytorch) (Recommended): The third version of our Automatic Speech Recognition and Subtitle Generation model. It is a fine-tuned version of ASR_subtitles_v2 without Kaldi-dependency (pure Pytorch), and refined training data leveraging contextualisation techniques for pseudo-labeling.
|
| 29 |
|
| 30 |
+
-- [**ASR_subtitles_v2**](https://huggingface.co/nelfproject/ASR_subtitles_v2): The second version of our Automatic Speech Recognition and Subtitle Generation model, with improved architecture and trained on 14000 hours of Flemish broadcast subtitled speech data.
|
| 31 |
It can generate both an exact verbatim transcription with annotation tags as well as a fully formatted and cleaned up subtitle transcription.
|
| 32 |
|
| 33 |
+
-- [**ASR_subtitles_v2_small**](https://huggingface.co/nelfproject/ASR_subtitles_v2_small): Smaller variant of ASR_subtitles_v2 with almost as good performance.
|
| 34 |
|
| 35 |
+
-- [**ASR_subtitles_v1**](https://huggingface.co/nelfproject/ASR_subtitles_v1): The first version of the ASR and Subtitling model trained on 1000 hours of Flemish data.
|
| 36 |
|
| 37 |
+
-- [**ASR_verbatim_v1**](https://huggingface.co/nelfproject/ASR_verbatim_v1): The first version of the ASR and Subtitling model trained on 1000 hours of Flemish data, converted to a verbatim-only ASR model.
|
| 38 |
|
| 39 |
-- **Whisper**: A finetuned Whisper Large model on Flemish data can be found [here](https://huggingface.co/kul-speech-lab/whisper_large_CGN). Usage instructions can be found in Whisper documentation.
|
| 40 |
|
| 41 |
+
**USAGE**: To use our ASR models and transcribe speech yourself, use [our Github](https://github.com/nelfproject/NeLF_Speech2Text_Pytorch) for NeLF_S2T_Pytorch or [our codebase](https://github.com/nelfproject/NeLF_Transcription_ASR) for previous versions.
|
| 42 |
|
| 43 |
### Speaker Diarization and Identification
|
| 44 |
|