| --- |
| configs: |
| - config_name: gsm8k_araeng |
| data_files: |
| - split: test |
| path: |
| - "gsm8k/gsm8k_araeng.csv" |
| - config_name: gsm8k_chieng |
| data_files: |
| - split: test |
| path: |
| - "gsm8k/gsm8k_chieng.csv" |
| - config_name: gsm8k_hineng |
| data_files: |
| - split: test |
| path: |
| - "gsm8k/gsm8k_hineng.csv" |
| - config_name: gsm8k_spaeng |
| data_files: |
| - split: test |
| path: |
| - "gsm8k/gsm8k_spaeng.csv" |
| - config_name: lid_chieng |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_chieng.csv" |
| - config_name: lid_fridut |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_fridut.csv" |
| - config_name: lid_gereng |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_gereng.csv" |
| - config_name: lid_guaspa |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_guaspa.csv" |
| - config_name: lid_hineng |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_hineng.csv" |
| - config_name: lid_hokman |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_hokman.csv" |
| - config_name: lid_mareng |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_mareng.csv" |
| - config_name: lid_msaea |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_msaea.csv" |
| - config_name: lid_nepeng |
| data_files: |
| - split: test |
| path: |
| - "lid/lid_nepeng.csv" |
| - config_name: mmlu_araeng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_araeng.csv" |
| - config_name: mmlu_beneng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_beneng.csv" |
| - config_name: mmlu_chieng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_chieng.csv" |
| - config_name: mmlu_duteng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_duteng.csv" |
| - config_name: mmlu_freeng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_freeng.csv" |
| - config_name: mmlu_gereng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_gereng.csv" |
| - config_name: mmlu_hineng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_hineng.csv" |
| - config_name: mmlu_mareng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_mareng.csv" |
| - config_name: mmlu_nepeng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_nepeng.csv" |
| - config_name: mmlu_spaeng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_spaeng.csv" |
| - config_name: mmlu_tameng |
| data_files: |
| - split: test |
| path: |
| - "mmlu/mmlu_tameng.csv" |
| - config_name: mt_araeng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_araeng_eng.csv" |
| - config_name: mt_beneng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_beneng_eng.csv" |
| - config_name: mt_chieng_chi |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_chieng_chi.csv" |
| - config_name: mt_chieng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_chieng_eng.csv" |
| - config_name: mt_hineng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_hineng_eng.csv" |
| - config_name: mt_hokman_man |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_hokman_man.csv" |
| - config_name: mt_mareng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_mareng_eng.csv" |
| - config_name: mt_spaeng_eng |
| data_files: |
| - split: test |
| path: |
| - "mt/mt_spaeng_eng.csv" |
| - config_name: ner_guaspa |
| data_files: |
| - split: test |
| path: |
| - "ner/ner_guaspa.csv" |
| - config_name: ner_hineng |
| data_files: |
| - split: test |
| path: |
| - "ner/ner_hineng.csv" |
| - config_name: ner_msaea |
| data_files: |
| - split: test |
| path: |
| - "ner/ner_msaea.csv" |
| - config_name: ner_spaeng |
| data_files: |
| - split: test |
| path: |
| - "ner/ner_spaeng.csv" |
| - config_name: pos_chieng |
| data_files: |
| - split: test |
| path: |
| - "pos/pos_chieng.csv" |
| - config_name: pos_fridut |
| data_files: |
| - split: test |
| path: |
| - "pos/pos_fridut.csv" |
| - config_name: pos_hineng |
| data_files: |
| - split: test |
| path: |
| - "pos/pos_hineng.csv" |
| - config_name: pos_spaeng |
| data_files: |
| - split: test |
| path: |
| - "pos/pos_spaeng.csv" |
| - config_name: sa_beneng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_beneng.csv" |
| - config_name: sa_hineng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_hineng.csv" |
| - config_name: sa_maleng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_maleng.csv" |
| - config_name: sa_mareng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_mareng.csv" |
| - config_name: sa_nepeng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_nepeng.csv" |
| - config_name: sa_spaeng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_spaeng.csv" |
| - config_name: sa_tameng |
| data_files: |
| - split: test |
| path: |
| - "sa/sa_tameng.csv" |
| - config_name: truthfulqa_araeng |
| data_files: |
| - split: test |
| path: |
| - "truthfulqa/truthfulqa_araeng.csv" |
| - config_name: truthfulqa_chieng |
| data_files: |
| - split: test |
| path: |
| - "truthfulqa/truthfulqa_chieng.csv" |
| - config_name: truthfulqa_hineng |
| data_files: |
| - split: test |
| path: |
| - "truthfulqa/truthfulqa_hineng.csv" |
| - config_name: truthfulqa_spaeng |
| data_files: |
| - split: test |
| path: |
| - "truthfulqa/truthfulqa_spaeng.csv" |
| license: apache-2.0 |
| language: |
| - zh |
| - en |
| - es |
| - hi |
| - de |
| - nl |
| - fy |
| - fr |
| - ar |
| - bn |
| - mr |
| - ne |
| - ta |
| - ml |
| - gn |
| - ne |
|
|
| size_categories: |
| - 10K<n<100K |
|
|
| task_categories: |
| - text-generation |
| - question-answering |
| - translation |
| - text-classification |
|
|
| tags: |
| - code-mixing |
| - multilingual |
| - llm-evaluation |
| - benchmark |
| --- |
| # ℹ️Dataset Card for CodeMixBench |
|
|
| ## [EMNLP'25] [CodeMixBench: Evaluating Code-Mixing Capabilities of LLMs Across 18 Languages](https://arxiv.org/abs/2507.18791) |
|
|
| <a href="https://github.com/Jeromeyluck/CodeMixBench" target="_blank"> |
| <img alt="Github" src="https://img.shields.io/badge/🐙-Github-blue" /> |
| </a> |
| |
| <a href="https://arxiv.org/abs/2507.18791" target="_blank"> |
| <img alt="Paper" src="https://img.shields.io/badge/📜-Paper-purple" /> |
| </a> |
| <a href="https://2025.emnlp.org/" target="_blank"> |
| <img alt="EMNLP 2025" src="https://img.shields.io/badge/Proceedings-EMNLP2025-blue" /> |
| </a> |
| |
|
|
|
|
| <!-- Provide a quick summary of the dataset. --> |
|
|
| Code-mixing is a linguistic phenomenon where multilingual speakers switch or mix two or more languages within a single utterance or conversation. |
| To evaluate LLMs’ comprehension of multilingual code-mixed texts, we introduce CodeMixBench, a benchmark comprising eight tasks across 18 languages. |
|
|
|  |
|
|
|
|
| ## 🔎Dataset Details |
|
|
| Our benchmark comprises synthesized datasets targeting knowledge reasoning, |
| mathematical reasoning, and truthfulness tasks, along with LID, POS, NER, SA, and MT tasks, |
| which have been adapted from open-source studies. |
|
|
|
|
| ### CodeMixBench vs. Others |
|
|
| Previous benchmarks, such as GLUECoS and LinCE, primarily focus on traditional NLP tasks and are limited to a small number of languages. |
| LinCE includes four language pairs and five NLP tasks: Language Identification(LID), |
| Part of Speech (POS), Named Entity Recognition (NER), Sentiment Analysis (SA), and Machine Translation (MT). |
| In contrast, GLUECoS covers two language pairs, lacks the MT task, but adds Question Answering (QA) and Natural Language Inference (NLI). |
| Our review of recent codemixing studies indicates that research extends beyond the language pairs used in LinCE and GLUECoS. |
| Therefore, we expanded to 16 language pairs and introduced tasks better suited for evaluating LLMs, |
| such as Multi-Choice, Math, and Truthfulness, resulting in a total of eight tasks. |
|
|
|  |
|
|
| ### Statistics of Synthetic Datasets |
| For knowledge reasoning, we developed the code-mixed MMLU (CM-MMLU) based on the MMLU test set, |
| featuring multiple-choice questions from 57 subjects to assess the model's comprehensive knowledge reasoning abilities. |
| For mathematical reasoning, we created the code-mixed GSM8K (CM-GSM8K), derived from the GSM8K test set, |
| which evaluates mathematical reasoning capabilities with each question including step-by-step solutions. |
| For truthfulness assessment, we constructed the code-mixed TruthfulQA (CM-TruthfulQA) using 817 multiple-choice |
| questions from the TruthfulQA test set. |
|
|
|  |
|
|
| ### Statistics of Collected Datasets |
| We selected and reconstructed 30 datasets from existing open-source projects. To comprehensively evaluate the performance of large |
| models on code-mixing, we aimed to encompass a diverse range of language families and tasks, prioritizing manually annotated datasets. |
| Ultimately, we cover traditional NLP tasks such as Language Identification (LID), Named Entity Recognition (NER), |
| Part-of-Speech tagging (POS), Sentiment Analysis(SA), and Machine Translation (MT), and cover 16 languages from seven language families: |
| Germanic(en, de, nl, fy), Sino-Tibetan (zh, hok), Romance(es), Afro-Asiatic (msa, ea), Indo-Aryan (hi, bn, ne,mr), Dravidian (ta, ml), and Tupian (gn). |
|
|
|  |
|
|
| ### Experience Results |
| We evaluate three families of LLMs on CodeMixBench, revealing consistent underperformance across all models on code-mixing |
| datasets involving language pairs from different language families. However, enhancements |
| in training data size, model scale, post-training, and few-shot learning can improve LLM performance on code-mixing datasets. |
|
|
|  |
|
|
|  |
|
|
|
|
|
|
| ## 🚀Load CodeMixBench |
|
|
| Taking the GSM8K task with mixed Chinese and English, gsm8k_chieng, as an example. |
| |
| ```python |
| from datasets import load_dataset |
|
|
| dataset_dict = load_dataset('CodeMixBench/CodeMixBench', data_files={'test': './gsm8k/gsm8k_chieng.csv'}) |
| ``` |
| |
| ### 📍Dataset Sources |
| |
| <!-- Provide the basic links for the dataset. --> |
| |
| - **Repository:** https://github.com/Jeromeyluck/CodeMixBench/ |
| - **Paper:** [CodeMixBench: Evaluating Code-Mixing Capabilities of LLMs Across 18 Languages](https://huggingface.co/papers/2507.18791) |
| |
| ## Setup |
| |
| 1. Follow these steps to set up your development environment: |
| ```bash |
| git clone git@github.com:Jeromeyluck/CodeMixBench.git |
| cd CodeMixBench |
|
|
| conda create -n CodeMixBench python=3.9 |
| conda activate CodeMixBench |
| pip install -r requirements.txt |
| ``` |
| |
| 2. To launch an llm for testing: |
| ```bash |
| python ./test_model.py \ |
| --dataset lid_guaspa \ |
| --expid lid_guaspa_all_0shot \ |
| --model gpt-3.5-turbo \ |
| --shot 5 \ |
| --api sk-********************* \ |
| --url https://**************** |
| ``` |
| - `dataset`: select the dataset (e.g., `lid_gereng`, `lid_spaeng`, `ner_hineng`). |
| - `expid`: define the ID of the test, the results file will be named after this ID. |
| - `model`: the model you test. The default model is `gpt-3.5-turbo`. |
| - `shot`: use for few-shot test (by default it will be `1`). |
| - `api`: API Key (default key will be `OPENAI_API_KEY` defined in system path). |
| - `url`: API function provider's URL. |
| |
|
|
| ## 🔗Citation |
|
|
| <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. --> |
|
|
| **BibTeX:** |
|
|
| ``` |
| @misc{yang2025codemixbenchevaluatingcodemixingcapabilities, |
| title={CodeMixBench: Evaluating Code-Mixing Capabilities of LLMs Across 18 Languages}, |
| author={Yilun Yang and Yekun Chai}, |
| year={2025}, |
| eprint={2507.18791}, |
| archivePrefix={arXiv}, |
| primaryClass={cs.CL}, |
| url={https://arxiv.org/abs/2507.18791}, |
| } |
| ``` |
|
|
|
|
|
|