|
Ran
|
Jobs
3
|
Files
121
|
Run time
1min
|
Badge
README BADGES
|
push
github
Fix Fast Gradient Clipping bias gradient calculation for three dim data (#751) Summary: Pull Request resolved: https://github.com/pytorch/opacus/pull/751 The bias grad calculation for three dim data was incorect. Let `G = g^Tg`, where `g`, of dimensions `Txd` be the per-sample activation gradient, where `T` is the number of tokens and `d` dimension. The per-sample gradient norm with respect to bias is `vec(G)^T vec(1)`, instead of the erroneous,`vec(G)^T vec(G)` before. This diff fixes it. Reviewed By: aparna-aketi, HuanyuZhang Differential Revision: D70823094 fbshipit-source-id: c1fe1dd7f
15 of 15 new or added lines in 2 files covered. (100.0%)
5299 of 6188 relevant lines covered (85.63%)
1.91 hits per line
| ID | Job ID | Ran | Files | Coverage | |
|---|---|---|---|---|---|
| 1 | run-1 - 14370790931.1 | 120 |
85.47 |
GitHub Action Run | |
| 2 | run-2 - 14370790931.2 | 120 |
85.46 |
GitHub Action Run | |
| 3 | run-3 - 14370790931.3 | 66 |
48.49 |
GitHub Action Run |
| Coverage | ∆ | File | Lines | Relevant | Covered | Missed | Hits/Line |
|---|