Hi,
As always, thank you for this amazing contribution.
I am taking prof.bowman's class and attempting to run BioBERT https://github.com/dmis-lab/biobert-pytorch and SciBERT https://github.com/allenai/scibert on jiant. One of our objectives for the course project is to run BioBERT and SciBERT on common NLU tasks.
As far as I understand, it should be possible to add any Transformers encoder model to jiant, but both of those models will probably require a bit of code?
I am sketching out what I'd have to do in jiant vs. plain transformers. Will using jiant create more overheads (by trying to support those models) than just following the standard fine-tuning processes in transformers?
Any suggestions and pointers will be helpful. Thank you!