Skip to content

Results with pretrain_data_fraction args #1066

@claravania

Description

@claravania

I'm trying to train MNLI in low-data regime (~750 examples) on RoBERTa, using the pretrain_data_fraction argument. I'm using default RoBERTa hyperparameter setup, but set the num epochs higher to allow the model trained longer. However, I got quite high performance 89.02 macro_avg, just slightly lower than using full data, 90.2 macro_avg. Could there possibly be some bugs?

Attaching log file, config file, and my command.

Command:
python main.py --config_file jiant/config/nli-roberta_conf.txt -o "run_name=run1, random_seed=123456"

nli-roberta_conf.txt
log.log

(changed .conf to _conf.txt since I can't upload the format to GitHub)

Metadata

Metadata

Assignees

No one assigned

    Labels

    jiant-v1-legacyRelevant to versions <= v1.3.2

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions