-
Notifications
You must be signed in to change notification settings - Fork 55
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multi GPU with PEFT on LLM #102
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I havnt given this a fair shake but in general what is the better way of doing this vs using subprocess :-D Any ideas?
Not sure, if this answers your question, but I plan to extend the ZenML core to serve automated capabilities for the creation of wrappers + making calls. https://zenml.atlassian.net/jira/software/c/projects/OSSK/boards/13?selectedIssue=OSSK-535 |
@avishniakov Personally I feel they are quite unstable and unreliable... a better way would to use the internal library and do this in code right? |
In theory, we can hack around the |
I somewhat heavily reworked how the preparation of the functions was done in this project. This is tightly coupled with the changes on the ZenML side. Looking forward to some conceptual feedback. There are definitely a few weak points:
|
To be merged after 0.58.3/0.59.0 is released |
This PR brings the multi-GPU DDP showcase to the PEFT training.
It is using the function helper from the core and the other changes are mostly related to Accelerate specifically.
ZenML core companion PR: zenml-io/zenml#2746 [ blocking merging of this one ]