-
Notifications
You must be signed in to change notification settings - Fork 165
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[QUESTION] How to perform inference on trained model? #106
Comments
@sleepingcat4 Am also looking for the same have you got anything on this. |
@Naziyashaik09 I've actually solved this problem. |
@sleepingcat4 Please can you share me the code or the notebook link. |
The main colab has examples of how to tune the model and make predictions (aka inference): https://colab.research.google.com/github/google-research/big_vision/blob/main/big_vision/configs/proj/paligemma/finetune_paligemma.ipynb. Is there something specific that is unclear? |
Can you please share on how you address the issue? I tried the one provided on the original code, but it is not working. We would appreciate if you could give some ideas or if possible share the code. Thank you. |
It's not difficult. Change the model and tokenizer paths to your trained |
@akolesnikoff actually no. Unfortunately, off-the boat it is not apparent we just need to change the model and tokenizer paths LOL! |
the dataset of fine-tuning paligemma is missing, and the longcap100 dataset is now inaccessible from kaggle. |
I have followed-thru a tutorial posted by Roboflow. While it was clear and helpful for learning PaliGemma, I had been struggling to figure out, how inference on the trained model (.npz) file can be done. Can someone provide some pointers and help?
https://colab.research.google.com/github/roboflow-ai/notebooks/blob/main/notebooks/how-to-finetune-paligemma-on-detection-dataset.ipynb?ref=blog.roboflow.com#scrollTo=TGDFTYVnY4zn
The text was updated successfully, but these errors were encountered: