Datasets:

Modalities:
Text
Formats:
parquet
Languages:
Portuguese
ArXiv:
Libraries:
Datasets
pandas
License:
Samoed commited on
Commit
ec3db2b
·
verified ·
1 Parent(s): 1261696

Add dataset card

Browse files
Files changed (1) hide show
  1. README.md +90 -4
README.md CHANGED
@@ -6,11 +6,14 @@ language:
6
  license: cc-by-sa-4.0
7
  multilinguality: monolingual
8
  source_datasets:
9
- - mteb/told-br
10
  task_categories:
11
  - text-classification
 
 
 
 
12
  task_ids:
13
- - multi-label-classification
14
  - sentiment-analysis
15
  - sentiment-scoring
16
  - sentiment-classification
@@ -63,7 +66,7 @@ tags:
63
  | Reference | https://paperswithcode.com/dataset/told-br |
64
 
65
  Source datasets:
66
- - [mteb/told-br](https://huggingface.co/datasets/mteb/told-br)
67
 
68
 
69
  ## How to evaluate on this task
@@ -142,7 +145,90 @@ desc_stats = task.metadata.descriptive_stats
142
  ```
143
 
144
  ```json
145
- {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
146
  ```
147
 
148
  </details>
 
6
  license: cc-by-sa-4.0
7
  multilinguality: monolingual
8
  source_datasets:
9
+ - mteb/BrazilianToxicTweetsClassification
10
  task_categories:
11
  - text-classification
12
+ - sentiment-analysis
13
+ - sentiment-scoring
14
+ - sentiment-classification
15
+ - hate-speech-detection
16
  task_ids:
 
17
  - sentiment-analysis
18
  - sentiment-scoring
19
  - sentiment-classification
 
66
  | Reference | https://paperswithcode.com/dataset/told-br |
67
 
68
  Source datasets:
69
+ - [mteb/BrazilianToxicTweetsClassification](https://huggingface.co/datasets/mteb/BrazilianToxicTweetsClassification)
70
 
71
 
72
  ## How to evaluate on this task
 
145
  ```
146
 
147
  ```json
148
+ {
149
+ "test": {
150
+ "num_samples": 2048,
151
+ "number_texts_intersect_with_train": 23,
152
+ "text_statistics": {
153
+ "total_text_length": 172708,
154
+ "min_text_length": 5,
155
+ "average_text_length": 84.330078125,
156
+ "max_text_length": 304,
157
+ "unique_texts": 2046
158
+ },
159
+ "image_statistics": null,
160
+ "label_statistics": {
161
+ "min_labels_per_text": 0,
162
+ "average_label_per_text": 0.57958984375,
163
+ "max_labels_per_text": 4,
164
+ "unique_labels": 7,
165
+ "labels": {
166
+ "obscene": {
167
+ "count": 653
168
+ },
169
+ "insult": {
170
+ "count": 430
171
+ },
172
+ "misogyny": {
173
+ "count": 46
174
+ },
175
+ "racism": {
176
+ "count": 13
177
+ },
178
+ "xenophobia": {
179
+ "count": 13
180
+ },
181
+ "homophobia": {
182
+ "count": 32
183
+ },
184
+ "None": {
185
+ "count": 1145
186
+ }
187
+ }
188
+ }
189
+ },
190
+ "train": {
191
+ "num_samples": 8192,
192
+ "number_texts_intersect_with_train": null,
193
+ "text_statistics": {
194
+ "total_text_length": 714281,
195
+ "min_text_length": 4,
196
+ "average_text_length": 87.1925048828125,
197
+ "max_text_length": 322,
198
+ "unique_texts": 8172
199
+ },
200
+ "image_statistics": null,
201
+ "label_statistics": {
202
+ "min_labels_per_text": 0,
203
+ "average_label_per_text": 0.5751953125,
204
+ "max_labels_per_text": 4,
205
+ "unique_labels": 7,
206
+ "labels": {
207
+ "None": {
208
+ "count": 4580
209
+ },
210
+ "obscene": {
211
+ "count": 2576
212
+ },
213
+ "insult": {
214
+ "count": 1700
215
+ },
216
+ "homophobia": {
217
+ "count": 139
218
+ },
219
+ "misogyny": {
220
+ "count": 179
221
+ },
222
+ "racism": {
223
+ "count": 54
224
+ },
225
+ "xenophobia": {
226
+ "count": 64
227
+ }
228
+ }
229
+ }
230
+ }
231
+ }
232
  ```
233
 
234
  </details>