-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Custom Finetuned Models? #32
Comments
Hi, If you want to add new models, then follow these steps:
After these steps, you should be able to load the model via python. |
I do not want to upload the model to HuggingFace due to the nature of the models being private(although perhaps one can have private models on HuggingFace?). However, if the method you stated would work, I believe that I could add support for loading models locally. I will likely explore adding this feature. Thanks for your insight. |
Yes. If you don't want to upload, a hack would be to put the Then you can do use it by:
For example |
Is same work for hivemind/gpt-j-6B-8bit model ? |
You could load it at 4-bit. For now 8-bit isn't supported. |
My PR #38 should close this issue. Will close when merged |
Looking at the code, it appears that when loading a model the code is loading preprocessed models that are uploaded to Huggingface and then checks that sha256 to make sure that it matches. It does not seem that the code currently allows loading a model from a path.
I have converted GPTJ myself into the GGML format(which I am almost certain this is built from, correct me if wrong).
I am interested in finetuning a model, converting the model to GGML, quantize to 4 bits, and then using the model through an API.
I believe that should this repo support custom fine-tuned models, it would be great for this use case.
The text was updated successfully, but these errors were encountered: