-
Notifications
You must be signed in to change notification settings - Fork 233
Top-K KL Divergence loss #747
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Signed-off-by: Asha Anoosheh <[email protected]>
Signed-off-by: Asha Anoosheh <[email protected]>
Signed-off-by: Asha Anoosheh <[email protected]>
Signed-off-by: Asha Anoosheh <[email protected]>
|
Auto-sync is disabled for draft pull requests in this repository. Workflows must be run manually. Contributors can view more details about this message here. |
Signed-off-by: Asha Anoosheh <[email protected]>
06d057f to
e7d33a7
Compare
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #747 +/- ##
==========================================
- Coverage 74.68% 74.62% -0.06%
==========================================
Files 192 192
Lines 18950 18989 +39
==========================================
+ Hits 14153 14171 +18
- Misses 4797 4818 +21 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
Signed-off-by: Asha Anoosheh <[email protected]>
Signed-off-by: Asha Anoosheh <[email protected]>
e7a34bf to
3093b8a
Compare
Signed-off-by: Asha Anoosheh <[email protected]>
What does this PR do?
Type of change: New feature
Overview: Writes a new KLDiv Logits loss which only uses top-k vocab values
Usage
# Add a code snippet demonstrating how to use thisTesting
Before your PR is "Ready for review"
Additional Information