Come on, that's not possible.
Were you meaning the V100 SXM?
No, we are trying to figure out if its just too early with pytorch or a driver issue or something else?
And by faster I ment lower latency per frame processed in small batch sizes, not more frames processed in big batches.
|===============================+======================+======================|
| 0 A100-SXM4-40GB On | 00000000:01:00.0 Off | 0 |
| N/A 29C P0 62W / 400W | 6776MiB / 40536MiB | 2% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 1 A100-SXM4-40GB On | 00000000:41:00.0 Off | 0 |
| N/A 30C P0 60W / 400W | 6716MiB / 40536MiB | 4% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 2 A100-SXM4-40GB On | 00000000:81:00.0 Off | 0 |
| N/A 26C P0 57W / 400W | 6576MiB / 40536MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+
| 3 A100-SXM4-40GB On | 00000000:C1:00.0 Off | 0 |
| N/A 27C P0 60W / 400W | 6648MiB / 40536MiB | 0% Default |
| | | Disabled |
+-------------------------------+----------------------+----------------------+