Update README.md
Browse files
README.md
CHANGED
|
@@ -11,8 +11,9 @@ text-to-speech model, whereas the grammatically incorrect texts come from the [C
|
|
| 11 |
|
| 12 |
## Introduction
|
| 13 |
|
| 14 |
-
The Synthesized English Speech with Grammatical Errors (SESGE) dataset was developed to support the [DeMINT](https://github.com/transducens/demint) project
|
| 15 |
-
|
|
|
|
| 16 |
feedback on the transcripts of their online meetings. As part of this, a system able to transcribe spoken English keeping the original
|
| 17 |
grammatical errors intact was essential.
|
| 18 |
Existing speech-to-text (STT) models like Whisper tend to correct grammatical errors due to their strong internal language models, making them unsuitable for this task.
|
|
@@ -22,15 +23,14 @@ Therefore, SESGE was created to train a custom STT model that could accurately t
|
|
| 22 |
|
| 23 |
Given the absence of a suitable dataset for training an error-preserving STT system, DeMINT fine-tuned a Whisper model with data from two primary sources:
|
| 24 |
|
| 25 |
-
- [COREFL](https://www.peterlang.com/document/1049094)
|
|
|
|
| 26 |
The COREFL dataset consists of essays written by non-native English students with various levels of proficiency.
|
| 27 |
While some of these essays have associated audio recordings, the majority do not.
|
| 28 |
To expand the audio dataset, we used the [StyleTTS2](https://arxiv.org/abs/2306.07691) text-to-speech model to generate synthetic audio for the remaining texts.
|
| 29 |
Multiple voices were used for synthesis to increase the diversity of the dataset.
|
| 30 |
-
COREFL also includes audio directly recorded by students, which introduces natural speech variability and common errors found among L1-Spanish speakers,
|
| 31 |
-
a key demographic for the DeMINT project.
|
| 32 |
|
| 33 |
-
- [C4_200M](https://github.com/google-research-datasets/C4_200M-synthetic-dataset-for-grammatical-error-correction)
|
| 34 |
The C4_200M dataset contains synthetically generated English sentences with grammatical errors, produced using a corruption model.
|
| 35 |
Like with COREFL, StyleTTS2 was employed to synthesize audio from these texts, diversifying the voices to enhance the training set.
|
| 36 |
This dataset primarily provides varied sentence structures and error types, although with a limited number of distinct voices.
|
|
@@ -38,14 +38,14 @@ This dataset primarily provides varied sentence structures and error types, alth
|
|
| 38 |
Due to licensing restrictions associated with the COREFL dataset, only the portion derived from the C4_200M dataset is publicly available as part of the
|
| 39 |
SESGE dataset. This means that while COREFL data was used during our training, only the C4_200M-based data is included in this dataset.
|
| 40 |
|
| 41 |
-
Training samples comprise
|
| 42 |
|
| 43 |
## Models
|
| 44 |
|
| 45 |
Two models were trained on the SESGE dataset by fine-tuning Whisper, enabling error-preserving STT. These models are available on the Hugging Face Hub:
|
| 46 |
|
| 47 |
-
- [Error-Preserving Whisper
|
| 48 |
-
- [Error-Preserving Whisper
|
| 49 |
|
| 50 |
Both models have been optimized to transcribe spoken English while retaining grammatical errors, making them suitable for language-learning applications
|
| 51 |
where fidelity to spoken errors is essential.
|
|
|
|
| 11 |
|
| 12 |
## Introduction
|
| 13 |
|
| 14 |
+
The Synthesized English Speech with Grammatical Errors (SESGE) dataset was developed to support the [DeMINT](https://github.com/transducens/demint) project
|
| 15 |
+
developed at Universitat d'Alacant, Spain.
|
| 16 |
+
The objective of DeMINT was to develop an intelligent tutoring system that helps non-native English speakers improve their language skills by analyzing and providing
|
| 17 |
feedback on the transcripts of their online meetings. As part of this, a system able to transcribe spoken English keeping the original
|
| 18 |
grammatical errors intact was essential.
|
| 19 |
Existing speech-to-text (STT) models like Whisper tend to correct grammatical errors due to their strong internal language models, making them unsuitable for this task.
|
|
|
|
| 23 |
|
| 24 |
Given the absence of a suitable dataset for training an error-preserving STT system, DeMINT fine-tuned a Whisper model with data from two primary sources:
|
| 25 |
|
| 26 |
+
- [COREFL](https://www.peterlang.com/document/1049094) (dataset [here](http://corefl.learnercorpora.com
|
| 27 |
+
)).
|
| 28 |
The COREFL dataset consists of essays written by non-native English students with various levels of proficiency.
|
| 29 |
While some of these essays have associated audio recordings, the majority do not.
|
| 30 |
To expand the audio dataset, we used the [StyleTTS2](https://arxiv.org/abs/2306.07691) text-to-speech model to generate synthetic audio for the remaining texts.
|
| 31 |
Multiple voices were used for synthesis to increase the diversity of the dataset.
|
|
|
|
|
|
|
| 32 |
|
| 33 |
+
- [C4_200M](https://github.com/google-research-datasets/C4_200M-synthetic-dataset-for-grammatical-error-correction).
|
| 34 |
The C4_200M dataset contains synthetically generated English sentences with grammatical errors, produced using a corruption model.
|
| 35 |
Like with COREFL, StyleTTS2 was employed to synthesize audio from these texts, diversifying the voices to enhance the training set.
|
| 36 |
This dataset primarily provides varied sentence structures and error types, although with a limited number of distinct voices.
|
|
|
|
| 38 |
Due to licensing restrictions associated with the COREFL dataset, only the portion derived from the C4_200M dataset is publicly available as part of the
|
| 39 |
SESGE dataset. This means that while COREFL data was used during our training, only the C4_200M-based data is included in this dataset.
|
| 40 |
|
| 41 |
+
Training samples comprise 28,592 utterances from C4_200M. Validation and test sets contain 700 samples each.
|
| 42 |
|
| 43 |
## Models
|
| 44 |
|
| 45 |
Two models were trained on the SESGE dataset by fine-tuning Whisper, enabling error-preserving STT. These models are available on the Hugging Face Hub:
|
| 46 |
|
| 47 |
+
- [Error-Preserving Whisper model](https://huggingface.co/Transducens/error-preserving-whisper)
|
| 48 |
+
- [Error-Preserving Whisper distilled model](https://huggingface.co/Transducens/error-preserving-whisper-distilled)
|
| 49 |
|
| 50 |
Both models have been optimized to transcribe spoken English while retaining grammatical errors, making them suitable for language-learning applications
|
| 51 |
where fidelity to spoken errors is essential.
|