-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Finetuning on FLUE #32
Comments
Ok, I went through this by writting "encoding = 'utf-8' on line 61 of extract_pawsx.py : Anyway, this could be usefull for future users of flue whith hugging face library given script. |
Hi @LisaBanana, Thanks a lot for your interest! I'm glad that you made it work for you. I'm going to update the code with your fix. Have a nice day! |
Hi @formiel ! Thanks for your message, actually I do have a new issue. So, I'm trying to finetune Flaubert on th MRPC Glue task, my command line is : From that point I opened the glue.py file and the run_glue.py files to see what could be done but it's kinda out of my reach. Update : And same for CoLA task : It seems to come from the train.tsv but the only thing I change on the data process is the encoding line from my previous question so I don't understand what I'm doing wrong. Have a nice day :) |
Hi @LisaBanana, Could you please try looking at the file
|
That’s strange… I’ve tried again the pipeline, with the latest version of transformer, and it works. I would suggest to clone the transformers repo and install it in editable mode ( |
It still doesn't work, I tried to reinstall everything in an other virtual env but still. Would you mind sharing a requirements.txt with the versions of your libraries / packages ? I really don't understand what's wrong with what I do ... |
Ok, thanks to my genius co-worker, we've got it ! The tsv file as it was processed (from the script you provide on the repo, which is strange btw if you don't have the same issues I encountered) had some unexpected "/n" and it was the reason why everything was brpken. Anyway, thanks for your help before :) |
Hi there, I'm a colleague of LisaBanana, and we managed to make the whole thing work! Unfortunately, it seems that the training will take 140 hours. Is this expected? ii'm using transformers 3.3.1 and running on a Tesla k80 GPU. Here is the command i'm using:
Have a good day, |
Hi @LisaBanana,
Sorry I was a little bit overwhelmed during the last few days and forgot to answer you. I'm glad that it worked for you. That's strange, the line break is Hi @AmauryLepicard, Thanks for your help to make the code run previously!
Oh I think that's not expected. However, I think your training should take around 19.31 hours instead of 140 hours (184260 steps / 2.65s / 3600)? |
Actually I'm not the "genius coder", he was another of our colleagues :-)
It would be 19h if it was 2.65 steps per second, but it is the opposite, 2.65 seconds per step! |
Hi @AmauryLepicard,
Oh sorry for my mistake. So thanks for your interest!
Oops, that's right. The code displays number of steps per second on my side, so I overlooked and thought it's the same case for you (maybe because I'm using transformers version 3.0.2). Training it on 2 Quadro P6000 GPUs with per-gpu batch size as yours takes me around 22 hours. Assuming the training time is scaled linearly, it would take me around 44 hours to train on 1 GPU. |
Hi !
I would like to finetune Flaubert on FLUE task with hugging face library. I downloaded the PAWS datas and use the code you gave on your github repo but I have this error message I can't go through :
any idea on what to do ?
Thanks for this project by the way, him looking forward to use it !
Have a good day,
Lisa
The text was updated successfully, but these errors were encountered: