Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

(not an issue) training 124m model works perfectly on 3060 #299

Closed
FlashlightET opened this issue Jul 29, 2022 · 0 comments
Closed

(not an issue) training 124m model works perfectly on 3060 #299

FlashlightET opened this issue Jul 29, 2022 · 0 comments

Comments

@FlashlightET
Copy link

i was worried i would run out of vram (i doubted it, since a 16gb cloud tpu could finetune both the 124m and 355m models, i wouldnt see why my 12gb gpu wouldnt be able to finetune the 124m model) but i'm going, surprisingly quick, with 10.1/12gb used up. i couldn't get aitextgen working though due to dumb python dependencies and idiotic tensorflow deprecations though. but i'm training 124m flawlessly. thank you for this tool

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant