gsaltintas commited on
Commit
f959648
·
verified ·
1 Parent(s): f94b238

Uploading tokenizer_robustness_completion_farsi_number_romanization subset

Browse files
README.md CHANGED
@@ -36,6 +36,10 @@ configs:
36
  data_files:
37
  - split: test
38
  path: tokenizer_robustness_completion_farsi_keyboard_proximity_errors/test-*
 
 
 
 
39
  dataset_info:
40
  - config_name: tokenizer_robustness_completion_farsi_arabic_keyboard_for_farsi
41
  features:
@@ -905,6 +909,130 @@ dataset_info:
905
  num_examples: 40
906
  download_size: 40791
907
  dataset_size: 23457
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
908
  ---
909
 
910
  # Dataset Card for Tokenization Robustness
 
36
  data_files:
37
  - split: test
38
  path: tokenizer_robustness_completion_farsi_keyboard_proximity_errors/test-*
39
+ - config_name: tokenizer_robustness_completion_farsi_number_romanization
40
+ data_files:
41
+ - split: test
42
+ path: tokenizer_robustness_completion_farsi_number_romanization/test-*
43
  dataset_info:
44
  - config_name: tokenizer_robustness_completion_farsi_arabic_keyboard_for_farsi
45
  features:
 
909
  num_examples: 40
910
  download_size: 40791
911
  dataset_size: 23457
912
+ - config_name: tokenizer_robustness_completion_farsi_number_romanization
913
+ features:
914
+ - name: question
915
+ dtype: string
916
+ - name: choices
917
+ list: string
918
+ - name: answer
919
+ dtype: int64
920
+ - name: answer_label
921
+ dtype: string
922
+ - name: split
923
+ dtype: string
924
+ - name: subcategories
925
+ dtype: string
926
+ - name: category
927
+ dtype: string
928
+ - name: lang
929
+ dtype: string
930
+ - name: second_lang
931
+ dtype: string
932
+ - name: notes
933
+ dtype: string
934
+ - name: id
935
+ dtype: string
936
+ - name: set_id
937
+ dtype: string
938
+ - name: variation_id
939
+ dtype: string
940
+ - name: vanilla_cos_sim_to_canonical
941
+ struct:
942
+ - name: CohereLabs/aya-expanse-8b
943
+ dtype: float64
944
+ - name: Qwen/Qwen3-8B
945
+ dtype: float64
946
+ - name: bigscience/bloom
947
+ dtype: float64
948
+ - name: common-pile/comma-v0.1-1t
949
+ dtype: float64
950
+ - name: facebook/xglm-564M
951
+ dtype: float64
952
+ - name: google-bert/bert-base-multilingual-cased
953
+ dtype: float64
954
+ - name: google/byt5-small
955
+ dtype: float64
956
+ - name: google/gemma-2-2b
957
+ dtype: float64
958
+ - name: gpt2
959
+ dtype: float64
960
+ - name: meta-llama/Llama-3.2-1B
961
+ dtype: float64
962
+ - name: microsoft/Phi-3-mini-4k-instruct
963
+ dtype: float64
964
+ - name: mistralai/tekken
965
+ dtype: float64
966
+ - name: tiktoken/gpt-4o
967
+ dtype: float64
968
+ - name: tokenmonster/englishcode-32000-consistent-v1
969
+ dtype: float64
970
+ - name: trimmed_cos_sim_to_canonical
971
+ struct:
972
+ - name: CohereLabs/aya-expanse-8b
973
+ dtype: float64
974
+ - name: Qwen/Qwen3-8B
975
+ dtype: float64
976
+ - name: bigscience/bloom
977
+ dtype: float64
978
+ - name: common-pile/comma-v0.1-1t
979
+ dtype: float64
980
+ - name: facebook/xglm-564M
981
+ dtype: float64
982
+ - name: google-bert/bert-base-multilingual-cased
983
+ dtype: float64
984
+ - name: google/byt5-small
985
+ dtype: float64
986
+ - name: google/gemma-2-2b
987
+ dtype: float64
988
+ - name: gpt2
989
+ dtype: float64
990
+ - name: meta-llama/Llama-3.2-1B
991
+ dtype: float64
992
+ - name: microsoft/Phi-3-mini-4k-instruct
993
+ dtype: float64
994
+ - name: mistralai/tekken
995
+ dtype: float64
996
+ - name: tiktoken/gpt-4o
997
+ dtype: float64
998
+ - name: tokenmonster/englishcode-32000-consistent-v1
999
+ dtype: float64
1000
+ - name: token_counts
1001
+ struct:
1002
+ - name: CohereLabs/aya-expanse-8b
1003
+ dtype: int64
1004
+ - name: Qwen/Qwen3-8B
1005
+ dtype: int64
1006
+ - name: bigscience/bloom
1007
+ dtype: int64
1008
+ - name: common-pile/comma-v0.1-1t
1009
+ dtype: int64
1010
+ - name: facebook/xglm-564M
1011
+ dtype: int64
1012
+ - name: google-bert/bert-base-multilingual-cased
1013
+ dtype: int64
1014
+ - name: google/byt5-small
1015
+ dtype: int64
1016
+ - name: google/gemma-2-2b
1017
+ dtype: int64
1018
+ - name: gpt2
1019
+ dtype: int64
1020
+ - name: meta-llama/Llama-3.2-1B
1021
+ dtype: int64
1022
+ - name: microsoft/Phi-3-mini-4k-instruct
1023
+ dtype: int64
1024
+ - name: mistralai/tekken
1025
+ dtype: int64
1026
+ - name: tiktoken/gpt-4o
1027
+ dtype: int64
1028
+ - name: tokenmonster/englishcode-32000-consistent-v1
1029
+ dtype: int64
1030
+ splits:
1031
+ - name: test
1032
+ num_bytes: 7500
1033
+ num_examples: 13
1034
+ download_size: 30813
1035
+ dataset_size: 7500
1036
  ---
1037
 
1038
  # Dataset Card for Tokenization Robustness
tokenizer_robustness_completion_farsi_number_romanization/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78969a6c77f53688d30c68d12fd9a727270f85aceda5ebe7da20be257c94d473
3
+ size 30813