We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
After distillation, the logs of training a student model is very large. you can find about model here https://huggingface.co/deepset/gelectra-large i am following distillation as following https://kd-lib.readthedocs.io/en/latest/usage/tutorials/VanillaKD.html
Epoch: 10, Loss: 2242.5531675815582, Accuracy: 0.32735324814532735 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 11, Loss: 2242.59808665514, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 12, Loss: 2242.5806131362915, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 13, Loss: 2242.5909256339073, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 14, Loss: 2242.556157052517, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 15, Loss: 2242.602696120739, Accuracy: 0.3274284693426608 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 16, Loss: 2242.567252278328, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 17, Loss: 2242.5414638519287, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 18, Loss: 2242.5438897013664, Accuracy: 0.3273814560943274 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 19, Loss: 2242.551034271717, Accuracy: 0.32739085874399404 -------------------------------------------------------------------------------- Validation Accuracy: 0.32513445409755914 Epoch: 20, Loss: 2242.530257344246, Accuracy: 0.3273814560943274
The text was updated successfully, but these errors were encountered:
No branches or pull requests
After distillation, the logs of training a student model is very large. you can find about model here
https://huggingface.co/deepset/gelectra-large
i am following distillation as following
https://kd-lib.readthedocs.io/en/latest/usage/tutorials/VanillaKD.html
The text was updated successfully, but these errors were encountered: