-
Notifications
You must be signed in to change notification settings - Fork 297
Closed
Labels
jiant-v1-legacyRelevant to versions <= v1.3.2Relevant to versions <= v1.3.2
Description
I'm trying to train MNLI in low-data regime (~750 examples) on RoBERTa, using the pretrain_data_fraction argument. I'm using default RoBERTa hyperparameter setup, but set the num epochs higher to allow the model trained longer. However, I got quite high performance 89.02 macro_avg, just slightly lower than using full data, 90.2 macro_avg. Could there possibly be some bugs?
Attaching log file, config file, and my command.
Command:
python main.py --config_file jiant/config/nli-roberta_conf.txt -o "run_name=run1, random_seed=123456"
(changed .conf to _conf.txt since I can't upload the format to GitHub)
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
jiant-v1-legacyRelevant to versions <= v1.3.2Relevant to versions <= v1.3.2