Automotive A100 SXM2 for FSD? (NVIDIA DRIVE A100)

Notice: Page may contain affiliate links for which we may earn a small commission through services like Amazon Affiliates or Skimlinks.

Leiko

New Member
Aug 15, 2021
14
0
1
On paper you can expect around 150TFlops in TF32 (18bit)
A100 SXM4 40G card has TF32 155TFlops


// The drive card has more rops over tesla A100 models, which should increase its performance processing images in int8 over normal A100 cards.
(Like resnet50 which typically runs with int8 precision -- non-image processing wise it might be bit slower than 40Gig SXM4 card)

This would be good comparison for AI workloads
View attachment 39905
(you can potentially expect int8, int4 and binary tops to be some 15-30% faster on the drive card -- i don't see them meaning quite a lot over 12bit precision)
where did you find the correct ROPs count ? The one on gpu database is wrong (the number of SMs is also wrong there)
 

CyklonDX

Well-Known Member
Nov 8, 2022
1,219
419
83
I must've read it wrong / clearly shows 128 rops and 192 tmu's.
1732633109883.png

must've mistaken it in my memory with this one here


// its even weaker than i remembered. *(i've got that gpu-z screen from some chinese site - don't have remember where anymore)
 
Last edited:

Leiko

New Member
Aug 15, 2021
14
0
1
I must've read it wrong / clearly shows 128 rops and 192 tmu's.
View attachment 40186

must've mistaken it in my memory with this one here


// its even weaker than i remembered. *(i've got that gpu-z screen from some chinese site - don't have remember where anymore)
I'm pretty sure you were correct about it being better for int8 than regular A1001732747759386.png1732747773723.png