Skip to content
This repository has been archived by the owner on Dec 24, 2024. It is now read-only.

Add N-distill #176

Open
3 of 4 tasks
pwuethri opened this issue Mar 11, 2019 · 1 comment
Open
3 of 4 tasks

Add N-distill #176

pwuethri opened this issue Mar 11, 2019 · 1 comment
Assignees

Comments

@pwuethri
Copy link
Contributor

pwuethri commented Mar 11, 2019

Adding N-distill according to https://arxiv.org/abs/1902.02186

  • Add next observation to trajectory data structure
  • Directly compute gradient using the given update rule (This is the difference compared to Teacher distill, on-policy distill and entropy regularised distillation
  • Update nn parameters accordingly
  • Test is policy distillation works using available teacher policy
@pwuethri pwuethri self-assigned this Mar 11, 2019
@pwuethri
Copy link
Contributor Author

Script exits with
terminate called after throwing an instance of 'std::runtime_error what(): invalid argument 13: ldc should be at least max(1, m=0), but have 0 at /pytorch/aten/src/TH/generic/THBlas.cpp:334 [1] 24572 abort (core dumped) python run_n_distill.py
-> need to check the computed gradients

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

1 participant