1. This site uses cookies. By continuing to use this site, you are agreeing to our use of cookies. Learn More.

Tesla P100 x 8 Linpack testing on SYS-4028GR-TXRT

Discussion in 'Linux-Bench Results and Discussion' started by dhenzjhen, Oct 20, 2017.

  1. dhenzjhen

    dhenzjhen Member

    Joined:
    Sep 14, 2016
    Messages:
    36
    Likes Received:
    53
    System: Supermicro SYS-4028GR-TXRT
    Motherboard: X10DG0-T
    CPU: E5 2699V4 x 2
    MEM: 32GB Micron x 12
    BIOS: 5/25/17
    GPU: Nvidia Tesla P100 SXM2 x 8
    OS: Ubuntu 16.04 x64
    Driver: 384.81
    CUDA: version 9


    [​IMG]
    [​IMG]
     
    #1
    Lukas Goe and Blinky 42 like this.
  2. PigLover

    PigLover Moderator

    Joined:
    Jan 26, 2011
    Messages:
    2,579
    Likes Received:
    1,004
    Now you are just showing off :):):)

    Seriously - impressive.
     
    #2
  3. dhenzjhen

    dhenzjhen Member

    Joined:
    Sep 14, 2016
    Messages:
    36
    Likes Received:
    53

    Tested P100 while am at it before switching the V100 tray :D
     
    #3
    Patrick likes this.
  4. Lukas Goe

    Lukas Goe New Member

    Joined:
    Jan 15, 2018
    Messages:
    1
    Likes Received:
    0
    Hello,

    I found your result here during my research and I am very interested in how exactly you achieved it. I have similar hardware, but my Gflops can't even get close to yours. I would be very grateful, if you could give me some additional informations or maybe even post your config files (HPL.dat) here.

    Which CUDA-Linpack version are you using? The only one I found and use seems rather old: hpl-2.0_FERMI_v15.

    I work on a cluster with 7 gpu nodes, each node got the following hardware:

    2x Intel Xeon E5-2640 v4
    8x DDR4-2400 8 GB Memory
    Intel X10DGQ Board
    4x Tesla P100 16GB HBM2

    CUDA 9.0

    The best I could achieve yet was roughly 3500 Gflops - with all 28 GPUs.. I think the benchmark isn't using the GPUs at all because nvidia-smi shows barely any usage (~45W/300W, 0% GPU-Util, ~2400 MiB Mem) and those 14 Xeons should be able to get close to 3500 Gflops on their own as far as I know. There is no warning or error whatsoever and everything always ends with PASSED.

    I would be very happy about any advice.
     
    #4
Similar Threads: Tesla P100
Forum Title Date
Linux-Bench Results and Discussion Tesla V100 x 8 Linpack testing on SYS-4028GR-TXRT Oct 6, 2017
Linux-Bench Results and Discussion Amber16 benchmark on 1028GQ-TXR and 4 P100SXM2 GPU Apr 2, 2017
Linux-Bench Results and Discussion GPU Linpack on 1028GQ-TXR and 4 P100SXM2 Mar 31, 2017
Linux-Bench Results and Discussion Gromacs benchmark on 1028GQ-TXR and 4 P100SXM2 GPU Mar 27, 2017

Share This Page