Skip to content

Latest commit

 

History

History
50 lines (35 loc) · 1.31 KB

README.md

File metadata and controls

50 lines (35 loc) · 1.31 KB

TrGLUE - A Natural Language Understanding Benchmark for Turkish

TrGLUE is a NLU benchmarking dataset for Turkish. As the name suggests, it's GLUE benchmarking dataset for Turkish language. You can download the datasets from the HuggingFace repo. For more information about the dataset, the tasks, data curation and more please visit the HF repo.

Benchmarking

Benchmarking code can be find under scripts/. To run a single task run run_single.sh:

#!/bin/bash

# Pick the task accordingly, here we pick COLA for an example

python3 run_trglue.py \
--model_name_or_path dbmdz/bert-base-turkish-cased \ 
--task_name cola \ 
--max_seq_length 128 \ 
--output_dir berturk \ 
--num_train_epochs 5 \ 
--learning_rate 2e-5 \ 
--per_device_train_batch_size 128 \ 
--per_device_eval_batch_size  128 \ 
--do_train \ 
--do_eval \ 
--do_predict

Available task names are:

  • cola
  • mnli
  • sst2
  • mrpc
  • qnli
  • qqp
  • rte
  • stsb
  • wnli

To run all the tasks in order, please run run_all.sh. Benchmarking for BERTurk model and a handful LLMs can be found under the HF repo and the research paper.

Research paper and citation

Coming soon!