Datasets:
Uploading tokenizer_robustness_completion_stem_scripted_text subset
Browse files
README.md
CHANGED
|
@@ -1372,6 +1372,130 @@ dataset_info:
|
|
| 1372 |
num_examples: 34
|
| 1373 |
download_size: 38084
|
| 1374 |
dataset_size: 18619
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1375 |
configs:
|
| 1376 |
- config_name: tokenizer_robustness_completion_stem_canonical
|
| 1377 |
data_files:
|
|
@@ -1417,6 +1541,10 @@ configs:
|
|
| 1417 |
data_files:
|
| 1418 |
- split: test
|
| 1419 |
path: tokenizer_robustness_completion_stem_morpheme_separation/test-*
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1420 |
---
|
| 1421 |
|
| 1422 |
# Dataset Card for Tokenization Robustness
|
|
|
|
| 1372 |
num_examples: 34
|
| 1373 |
download_size: 38084
|
| 1374 |
dataset_size: 18619
|
| 1375 |
+
- config_name: tokenizer_robustness_completion_stem_scripted_text
|
| 1376 |
+
features:
|
| 1377 |
+
- name: question
|
| 1378 |
+
dtype: string
|
| 1379 |
+
- name: choices
|
| 1380 |
+
list: string
|
| 1381 |
+
- name: answer
|
| 1382 |
+
dtype: int64
|
| 1383 |
+
- name: answer_label
|
| 1384 |
+
dtype: string
|
| 1385 |
+
- name: split
|
| 1386 |
+
dtype: string
|
| 1387 |
+
- name: subcategories
|
| 1388 |
+
dtype: string
|
| 1389 |
+
- name: lang
|
| 1390 |
+
dtype: string
|
| 1391 |
+
- name: second_lang
|
| 1392 |
+
dtype: string
|
| 1393 |
+
- name: notes
|
| 1394 |
+
dtype: string
|
| 1395 |
+
- name: id
|
| 1396 |
+
dtype: string
|
| 1397 |
+
- name: set_id
|
| 1398 |
+
dtype: string
|
| 1399 |
+
- name: variation_id
|
| 1400 |
+
dtype: string
|
| 1401 |
+
- name: question_general_category
|
| 1402 |
+
dtype: string
|
| 1403 |
+
- name: vanilla_cos_sim_to_canonical
|
| 1404 |
+
struct:
|
| 1405 |
+
- name: CohereLabs/aya-expanse-8b
|
| 1406 |
+
dtype: float64
|
| 1407 |
+
- name: Qwen/Qwen3-8B
|
| 1408 |
+
dtype: float64
|
| 1409 |
+
- name: bigscience/bloom
|
| 1410 |
+
dtype: float64
|
| 1411 |
+
- name: common-pile/comma-v0.1-1t
|
| 1412 |
+
dtype: float64
|
| 1413 |
+
- name: facebook/xglm-564M
|
| 1414 |
+
dtype: float64
|
| 1415 |
+
- name: google-bert/bert-base-multilingual-cased
|
| 1416 |
+
dtype: float64
|
| 1417 |
+
- name: google/byt5-small
|
| 1418 |
+
dtype: float64
|
| 1419 |
+
- name: google/gemma-2-2b
|
| 1420 |
+
dtype: float64
|
| 1421 |
+
- name: gpt2
|
| 1422 |
+
dtype: float64
|
| 1423 |
+
- name: meta-llama/Llama-3.2-1B
|
| 1424 |
+
dtype: float64
|
| 1425 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
| 1426 |
+
dtype: float64
|
| 1427 |
+
- name: mistralai/tekken
|
| 1428 |
+
dtype: float64
|
| 1429 |
+
- name: tiktoken/gpt-4o
|
| 1430 |
+
dtype: float64
|
| 1431 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
| 1432 |
+
dtype: float64
|
| 1433 |
+
- name: trimmed_cos_sim_to_canonical
|
| 1434 |
+
struct:
|
| 1435 |
+
- name: CohereLabs/aya-expanse-8b
|
| 1436 |
+
dtype: float64
|
| 1437 |
+
- name: Qwen/Qwen3-8B
|
| 1438 |
+
dtype: float64
|
| 1439 |
+
- name: bigscience/bloom
|
| 1440 |
+
dtype: float64
|
| 1441 |
+
- name: common-pile/comma-v0.1-1t
|
| 1442 |
+
dtype: float64
|
| 1443 |
+
- name: facebook/xglm-564M
|
| 1444 |
+
dtype: float64
|
| 1445 |
+
- name: google-bert/bert-base-multilingual-cased
|
| 1446 |
+
dtype: float64
|
| 1447 |
+
- name: google/byt5-small
|
| 1448 |
+
dtype: float64
|
| 1449 |
+
- name: google/gemma-2-2b
|
| 1450 |
+
dtype: float64
|
| 1451 |
+
- name: gpt2
|
| 1452 |
+
dtype: float64
|
| 1453 |
+
- name: meta-llama/Llama-3.2-1B
|
| 1454 |
+
dtype: float64
|
| 1455 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
| 1456 |
+
dtype: float64
|
| 1457 |
+
- name: mistralai/tekken
|
| 1458 |
+
dtype: float64
|
| 1459 |
+
- name: tiktoken/gpt-4o
|
| 1460 |
+
dtype: float64
|
| 1461 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
| 1462 |
+
dtype: float64
|
| 1463 |
+
- name: token_counts
|
| 1464 |
+
struct:
|
| 1465 |
+
- name: CohereLabs/aya-expanse-8b
|
| 1466 |
+
dtype: int64
|
| 1467 |
+
- name: Qwen/Qwen3-8B
|
| 1468 |
+
dtype: int64
|
| 1469 |
+
- name: bigscience/bloom
|
| 1470 |
+
dtype: int64
|
| 1471 |
+
- name: common-pile/comma-v0.1-1t
|
| 1472 |
+
dtype: int64
|
| 1473 |
+
- name: facebook/xglm-564M
|
| 1474 |
+
dtype: int64
|
| 1475 |
+
- name: google-bert/bert-base-multilingual-cased
|
| 1476 |
+
dtype: int64
|
| 1477 |
+
- name: google/byt5-small
|
| 1478 |
+
dtype: int64
|
| 1479 |
+
- name: google/gemma-2-2b
|
| 1480 |
+
dtype: int64
|
| 1481 |
+
- name: gpt2
|
| 1482 |
+
dtype: int64
|
| 1483 |
+
- name: meta-llama/Llama-3.2-1B
|
| 1484 |
+
dtype: int64
|
| 1485 |
+
- name: microsoft/Phi-3-mini-4k-instruct
|
| 1486 |
+
dtype: int64
|
| 1487 |
+
- name: mistralai/tekken
|
| 1488 |
+
dtype: int64
|
| 1489 |
+
- name: tiktoken/gpt-4o
|
| 1490 |
+
dtype: int64
|
| 1491 |
+
- name: tokenmonster/englishcode-32000-consistent-v1
|
| 1492 |
+
dtype: int64
|
| 1493 |
+
splits:
|
| 1494 |
+
- name: test
|
| 1495 |
+
num_bytes: 11267
|
| 1496 |
+
num_examples: 18
|
| 1497 |
+
download_size: 31156
|
| 1498 |
+
dataset_size: 11267
|
| 1499 |
configs:
|
| 1500 |
- config_name: tokenizer_robustness_completion_stem_canonical
|
| 1501 |
data_files:
|
|
|
|
| 1541 |
data_files:
|
| 1542 |
- split: test
|
| 1543 |
path: tokenizer_robustness_completion_stem_morpheme_separation/test-*
|
| 1544 |
+
- config_name: tokenizer_robustness_completion_stem_scripted_text
|
| 1545 |
+
data_files:
|
| 1546 |
+
- split: test
|
| 1547 |
+
path: tokenizer_robustness_completion_stem_scripted_text/test-*
|
| 1548 |
---
|
| 1549 |
|
| 1550 |
# Dataset Card for Tokenization Robustness
|
tokenizer_robustness_completion_stem_scripted_text/test-00000-of-00001.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98bb30c5bc9372cfaf7c356c1cbd29e1463013afd81352ee84619f75d1d62d78
|
| 3 |
+
size 31156
|