Skip to content

Commit 94caf02

Browse files
author
root
committed
Add more details of FP16 training/inference in model efficiency part in README.
1 parent 6a0f4e5 commit 94caf02

File tree

1 file changed

+10
-5
lines changed

1 file changed

+10
-5
lines changed

README.md

Lines changed: 10 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -175,11 +175,16 @@ Our BiRefNet has achieved SOTA on many similar HR tasks:
175175
> The devices used in the below table differ from those in the original paper (the standard). So, it's only for reference.
176176
177177
| Runtime | *FP32* | *FP16* |
178-
| :----: | :----: | :----: |
179-
| GPUMem | 4.76GB | 3.45GB |
180-
| A100 | 86.8ms | 69.4ms |
181-
| 4090 | 95.8ms | 57.7ms |
182-
| V100 | 384ms | 152ms |
178+
| :-----: | :----: | :----: |
179+
| A100 | 86.8ms | 69.4ms |
180+
| 4090 | 95.8ms | 57.7ms |
181+
| V100 | 384ms | 152ms |
182+
183+
| GPU Memory | *FP32* | *FP16* |
184+
| :--------: | :----: | :----: |
185+
| Inference | 4.76GB | 3.45GB |
186+
| Training (\#GPU=1, batch\_size=2, compile=False+PyTorch=2.5.1) | 36.3GB | 30.4GB |
187+
| Training (\#GPU=1, batch\_size=2, compile=True+PyTorch=2.5.1) | 35.9GB | **24.9GB** |
183188

184189
</details>
185190

0 commit comments

Comments
 (0)