Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LoRA beseline parameter #4

Open
jypppppp opened this issue Jan 10, 2024 · 2 comments
Open

LoRA beseline parameter #4

jypppppp opened this issue Jan 10, 2024 · 2 comments

Comments

@jypppppp
Copy link

Hi,

Thanks for your good work!

Can you clarify what is learning rate,bsz and epochs for baseline LoRA experiments among different datasets

Kind regards,

Jason

@telxt
Copy link
Collaborator

telxt commented Jan 11, 2024

Thank you for your interest in our work! The hyper-parameters of LoRA is listed in the following:

Dataset lr epoch
CoLA 8e-5 20
SST-2 1e-4 10
MRPC 1e-4 20
QQP 3e-4 10
STS-B 1e-4 20
MNLI 3e-4 10
QNLI 3e-4 10
RTE 1.2e-3 50

And the seed list is {0, 21, 42, 81, 100}, the batch_size is 8.

I hope my response helps you.

@ouxinwei111
Copy link

Hi, I was wondering if you use the same learning rate for all the rank settings. Looking forward to your help :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants