1.6TB intel NVMe ssd (p3605) (sun branded) ($280 usd)

Discussion in 'Great Deals' started by mineblaster, Feb 2, 2019.

  1. mineblaster

    mineblaster New Member

    Joined:
    Jan 16, 2019
    Messages:
    16
    Likes Received:
    1
    #1
  2. Oddworld

    Oddworld New Member

    Joined:
    Jan 16, 2018
    Messages:
    27
    Likes Received:
    0
    Anyone know if they can be flashed to default Intel p3600 firmware? If not, any limitations to be concerned about?

    I have the intel p3600 1.6TB and am curious whether they would be compatible, or whether firmware variations would prohibit RAID.
     
    #2
  3. T_Minus

    T_Minus Moderator

    Joined:
    Feb 15, 2015
    Messages:
    6,563
    Likes Received:
    1,362
    I run them together without problems p3605 and p3600
     
    #3
    gigatexal likes this.
  4. WeightedCube

    WeightedCube New Member

    Joined:
    Mar 11, 2015
    Messages:
    4
    Likes Received:
    2
    Well I bought 4 @ $200 each. Time to get some adapters to actually plug it in.
     
    #4
    gigatexal likes this.
  5. gigatexal

    gigatexal I'm here to learn

    Joined:
    Nov 25, 2012
    Messages:
    2,602
    Likes Received:
    471
    With the Sun firmwares?
     
    #5
  6. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    T_minus / Odd world - what kind / how are you using RAID with these NVMe ? (i assume some SW raid or NAS os) - if so how does RAID perofmrance look?

    im going to get 1 or 2x . fyi the seller has qty 42x left currently. tks for post!

    EDIT: why cant i find any specs / info on the intel p3605 ? i can only find on p3600 or p3610. is the p3605 oem only model maybe? tks
     
    #6
    mineblaster likes this.
  7. T_Minus

    T_Minus Moderator

    Joined:
    Feb 15, 2015
    Messages:
    6,563
    Likes Received:
    1,362
    I haven't benchmarked these specific drives in years... using them in ZFS though.

    When I test these I'll see if I can get them in a test system with 4x and 8x at once, I need to order some new AOC though to connect them at once so it'll be a couple weeks. They'll be paired with Optane.
     
    #7
  8. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    thanks tminus, are you using the nvme 's in raid currently or no? (if not , how are you using them?- just curuios).

    Im thinking im going to use the 4x 2.5" nvme's i have as host local, esxi datastores (1 in each host).
     
    #8
  9. T_Minus

    T_Minus Moderator

    Joined:
    Feb 15, 2015
    Messages:
    6,563
    Likes Received:
    1,362
    In ZFS pool of mirrors. Also in workstation PC as the 'work' drive for video\photo work.
     
    #9
    james23 likes this.
  10. WeightedCube

    WeightedCube New Member

    Joined:
    Mar 11, 2015
    Messages:
    4
    Likes Received:
    2
    #10
    james23 likes this.
  11. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    tks weighted cube.

    btw- i got 2x from this ebay seller at 200$ each (he accepted my 1st offer, i maybe could have gone lower).

    T_minus- now u got me going, im guessing you have 4x nvme's in a pool of of mirrors? is that a nas box or is it a box you work directly on? (i ask as: if its a nas box, are you getting full 10g speeds/file-transfers out of it to a client device?)
    i know this is all a bit off topic, but im pretty heavy into freenas/zfs these past few months (and going FW).
    tks
     
    #11
  12. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    i got my 2x in, the smart data are just about the same abt 2% used up (not bad for 200$!!). as expected you can NOT update these drives with normal intel firmware (so you prob cant update FW at all)

    will update later with nvme-cli data , this is from smartctl:

    Code:
    smartctl 6.6 2016-05-31 r4324 [x86_64-linux-4.15.0-29-generic] (local build)
    Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
    
    === START OF INFORMATION SECTION ===
    Model Number:                       INTEL SSDPE2ME016T4S
    Serial Number:                      CVMD505500EL1P6LGN
    Firmware Version:                   8DV1RA13
    PCI Vendor ID:                      0x8086
    PCI Vendor Subsystem ID:            0x108e
    IEEE OUI Identifier:                0x5cd2e4
    Controller ID:                      0
    Number of Namespaces:               1
    Namespace 1 Size/Capacity:          1,600,321,314,816 [1.60 TB]
    Namespace 1 Formatted LBA Size:     512
    Local Time is:                      Sat Feb 16 02:54:54 2019 UTC
    Firmware Updates (0x02):            1 Slot
    Optional Admin Commands (0x0006):   Format Frmw_DL
    Optional NVM Commands (0x0006):     Wr_Unc DS_Mngmt
    Maximum Data Transfer Size:         32 Pages
    
    Supported Power States
    St Op     Max   Active     Idle   RL RT WL WT  Ent_Lat  Ex_Lat
     0 +    25.00W       -        -    0  0  0  0        0       0
    
    Supported LBA Sizes (NSID 0x1)
    Id Fmt  Data  Metadt  Rel_Perf
     0 +     512       0         2
     1 -     512       8         2
     2 -     512      16         2
     3 -    4096       0         0
     4 -    4096       8         0
     5 -    4096      64         0
     6 -    4096     128         0
    
    === START OF SMART DATA SECTION ===
    SMART overall-health self-assessment test result: PASSED
    
    SMART/Health Information (NVMe Log 0x02, NSID 0x1)
    Critical Warning:                   0x00
    Temperature:                        22 Celsius
    Available Spare:                    100%
    Available Spare Threshold:          10%
    Percentage Used:                    2%
    Data Units Read:                    3,142,263 [1.60 TB]
    Data Units Written:                 316,682 [162 GB]
    Host Read Commands:                 44,451,731,018
    Host Write Commands:                15,363,551,223
    Controller Busy Time:               1,164
    Power Cycles:                       81
    Power On Hours:                     28,281
    Unsafe Shutdowns:                   73
    Media and Data Integrity Errors:    0
    Error Information Log Entries:      0
    
    Error Information (NVMe Log 0x01, max 64 entries)
    No Errors Logged
    
    

    Code:
    root@ubuntu:/home/ubuntu/intl# isdct show -intelssd
    
    - Intel SSD DC P3600 Series CVMD505500EL1P6LGN -
    
    Bootloader : 8B1B012E
    DevicePath : /dev/nvme0n1
    DeviceStatus : Healthy
    Firmware : 8DV1RA13
    FirmwareUpdateAvailable : Please contact your Intel representative about firmware update for this drive.
    Index : 0
    ModelNumber : INTEL SSDPE2ME016T4S
    ProductFamily : Intel SSD DC P3600 Series
    SerialNumber : CVMD505500EL1P6LGN
    
    
    
    edit: nvme-cli smart output for the same drives + it also appears to support 4k SS (and other SS), i have not tested this yet, but here is some output:

    Code:
    nvme-cli smart:
    Smart Log for NVME device:nvme0 namespace-id:ffffffff
    critical_warning                    : 0
    temperature                         : 22 C
    available_spare                     : 100%
    available_spare_threshold           : 10%
    percentage_used                     : 2%
    data_units_read                     : 3142263
    data_units_written                  : 316682
    host_read_commands                  : 44451731018
    host_write_commands                 : 15363551223
    controller_busy_time                : 1164
    power_cycles                        : 81
    power_on_hours                      : 28281
    unsafe_shutdowns                    : 73
    media_errors                        : 0
    num_err_log_entries                 : 0
    Warning Temperature Time            : 0
    Critical Composite Temperature Time : 0
    Thermal Management T1 Trans Count   : 0
    Thermal Management T2 Trans Count   : 0
    Thermal Management T1 Total Time    : 0
    Thermal Management T2 Total Time    : 0
    
    
    and multi sector sizes:
    
    NVME Identify Namespace 1:
    nsze    : 0xba4d4ab0
    ncap    : 0xba4d4ab0
    nuse    : 0xba4d4ab0
    nsfeat  : 0
    nlbaf   : 6
    flbas   : 0
    mc      : 0x1
    dpc     : 0x11
    dps     : 0
    nmic    : 0
    rescap  : 0
    fpi     : 0
    nawun   : 0
    nawupf  : 0
    nacwu   : 0
    nabsn   : 0
    nabo    : 0
    nabspf  : 0
    noiob   : 0
    nvmcap  : 0
    nguid   : 00000000000000000000000000000000
    eui64   : 0000000000000000
    lbaf  0 : ms:0   lbads:9  rp:0x2 (in use)
    lbaf  1 : ms:8   lbads:9  rp:0x2
    lbaf  2 : ms:16  lbads:9  rp:0x2
    lbaf  3 : ms:0   lbads:12 rp:0
    lbaf  4 : ms:8   lbads:12 rp:0
    lbaf  5 : ms:64  lbads:12 rp:0
    lbaf  6 : ms:128 lbads:12 rp:0 
     
    #12
    Last edited: Feb 15, 2019
    gigatexal likes this.
  13. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    for anyone interested, very strong performance on these quick tests i often run on nvmes. just as soemone said in that other older thread on these drive types, the oracle specs do seem to be very conservative / under spec'd! if i had fewer empty/unused 2.5" nvme's i would def buy more.

    note, the drive / tests are NOT Steady Stated , and most are run ontop of an ext4 format/mount (only some are direct to /dev/nvme.. as u can see) , all under ubuntu 18 live. (i have found in the past that i often get a bit better FIO numbers on nvme s, on an actual ubuntu 18 full install to a disk vs the live usb boot version, but these are under 18 live):

    -- (im seeing the performance is only a bit behind a 960gb hgst nvme - HUSMR7696BDP3Y1 drive , which is good @ this price too , as i just bought that 960g hgst ebay used at 189$ !)

    Code:
    root@ubuntu:~# fio --output=INTEL_fio_run_result.txt --name=myjob --filename=/dev/nvme0n1 --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=4K --rw=randread --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s
    Jobs: 4 (f=4): [r(4)][10.7%][r=2726MiB/s,w=0KiB/s][r=698k,w=0 IOPS][eta 04m:28s]
    Jobs: 4 (f=4): [r(4)][21.0%][r=2722MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 03m:57s]
    Jobs: 4 (f=4): [r(4)][31.3%][r=2724MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 03m:26s]
    Jobs: 4 (f=4): [r(4)][41.7%][r=2723MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 02m:55s]
    Jobs: 4 (f=4): [r(4)][52.0%][r=2721MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 02m:24s]
    Jobs: 4 (f=4): [r(4)][62.3%][r=2719MiB/s,w=0KiB/s][r=696k,w=0 IOPS][eta 01m:53s]
    Jobs: 4 (f=4): [r(4)][72.7%][r=2723MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 01m:22s]
    Jobs: 4 (f=4): [r(4)][83.0%][r=2726MiB/s,w=0KiB/s][r=698k,w=0 IOPS][eta 00m:51s]
    Jobs: 4 (f=4): [r(4)][93.3%][r=2729MiB/s,w=0KiB/s][r=699k,w=0 IOPS][eta 00m:20s]
    root@ubuntu:~# [r(4)][100.0%][r=2724MiB/s,w=0KiB/s][r=697k,w=0 IOPS][eta 00m:00s]
    
    
    
    
    root@ubuntu:~# fio --name=myjob --filename=/dev/nvme0n1 --direct=1 --norandommap --randrepeat=0 --runtime=60 --blocksize=4K --rw=randwrite --iodepth=32 --numjobs=2 --group_reporting --eta-newline=10s --name=mybs --size=4g --ioengine=libaio
    myjob: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=psync, iodepth=32
    ...
    mybs: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1
    fio-3.1
    Starting 3 processes
    mybs: Laying out IO file (1 file / 4096MiB)
    Jobs: 2 (f=2): [w(2),_(1)][19.4%][r=0KiB/s,w=516MiB/s][r=0,w=132k IOPS][eta 00m:50s]
    Jobs: 2 (f=2): [w(2),_(1)][37.1%][r=0KiB/s,w=476MiB/s][r=0,w=122k IOPS][eta 00m:39s]
    Jobs: 2 (f=2): [w(2),_(1)][54.8%][r=0KiB/s,w=515MiB/s][r=0,w=132k IOPS][eta 00m:28s]
    Jobs: 2 (f=2): [w(2),_(1)][72.6%][r=0KiB/s,w=496MiB/s][r=0,w=127k IOPS][eta 00m:17s]
    Jobs: 2 (f=2): [w(2),_(1)][90.3%][r=0KiB/s,w=487MiB/s][r=0,w=125k IOPS][eta 00m:06s]
    Jobs: 2 (f=2): [w(2),_(1)][96.8%][r=0KiB/s,w=517MiB/s][r=0,w=132k IOPS][eta 00m:02s]
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=4K --rw=randread --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=40g
    mybs: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    Jobs: 4 (f=4): [r(4)][38.1%][r=1958MiB/s,w=0KiB/s][r=501k,w=0 IOPS][eta 00m:52s]
    Jobs: 4 (f=4): [r(4)][75.9%][r=1924MiB/s,w=0KiB/s][r=492k,w=0 IOPS][eta 00m:20s]
    Jobs: 4 (f=4): [r(4)][100.0%][r=1962MiB/s,w=0KiB/s][r=502k,w=0 IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21002: Sat Feb 16 10:30:21 2019
       read: IOPS=501k, BW=1958MiB/s (2053MB/s)(160GiB/83673msec)
        slat (nsec): min=1522, max=12448k, avg=5493.56, stdev=6607.76
        clat (nsec): min=493, max=20346k, avg=249011.17, stdev=117496.18
         lat (usec): min=2, max=20350, avg=254.64, stdev=117.83
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=4K --rw=randwrite --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=40g
    
    mybs: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    mybs: Laying out IO file (1 file / 40960MiB)
    Jobs: 4 (f=4): [w(4)][11.0%][r=0KiB/s,w=230MiB/s][r=0,w=58.8k IOPS][eta 04m:28s]
    Jobs: 4 (f=4): [w(4)][21.3%][r=0KiB/s,w=264MiB/s][r=0,w=67.5k IOPS][eta 03m:57s]
    Jobs: 4 (f=4): [w(4)][31.6%][r=0KiB/s,w=283MiB/s][r=0,w=72.4k IOPS][eta 03m:26s]
    Jobs: 4 (f=4): [w(4)][41.9%][r=0KiB/s,w=323MiB/s][r=0,w=82.8k IOPS][eta 02m:55s]
    Jobs: 4 (f=4): [w(4)][52.2%][r=0KiB/s,w=364MiB/s][r=0,w=93.3k IOPS][eta 02m:24s]
    Jobs: 4 (f=4): [w(4)][62.5%][r=0KiB/s,w=388MiB/s][r=0,w=99.3k IOPS][eta 01m:53s]
    Jobs: 4 (f=4): [w(4)][72.8%][r=0KiB/s,w=468MiB/s][r=0,w=120k IOPS][eta 01m:22s]
    Jobs: 4 (f=4): [w(4)][83.1%][r=0KiB/s,w=621MiB/s][r=0,w=159k IOPS][eta 00m:51s]
    Jobs: 4 (f=4): [w(4)][93.4%][r=0KiB/s,w=702MiB/s][r=0,w=180k IOPS][eta 00m:20s]
    Jobs: 4 (f=4): [w(4)][100.0%][r=0KiB/s,w=740MiB/s][r=0,w=189k IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=20895: Sat Feb 16 10:23:07 2019
      write: IOPS=101k, BW=396MiB/s (415MB/s)(116GiB/300001msec)
        slat (usec): min=2, max=105688, avg=36.86, stdev=235.87
        clat (nsec): min=769, max=110568k, avg=1225235.14, stdev=1502863.56
         lat (usec): min=11, max=110601, avg=1262.31, stdev=1528.76
     
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=128K --rw=randwrite --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=40g
    mybs: (g=0): rw=randwrite, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    Jobs: 4 (f=4): [w(4)][30.8%][r=0KiB/s,w=1594MiB/s][r=0,w=12.7k IOPS][eta 01m:12s]
    Jobs: 4 (f=4): [w(4)][61.2%][r=0KiB/s,w=1585MiB/s][r=0,w=12.7k IOPS][eta 00m:40s]
    Jobs: 4 (f=4): [w(4)][91.3%][r=0KiB/s,w=1565MiB/s][r=0,w=12.5k IOPS][eta 00m:09s]
    Jobs: 4 (f=4): [w(4)][100.0%][r=0KiB/s,w=1574MiB/s][r=0,w=12.6k IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21169: Sat Feb 16 10:43:07 2019
      write: IOPS=12.7k, BW=1586MiB/s (1663MB/s)(160GiB/103306msec)
        slat (usec): min=9, max=35945, avg=40.40, stdev=140.87
        clat (usec): min=36, max=64328, avg=10040.61, stdev=5819.81
         lat (usec): min=64, max=64387, avg=10081.51, stdev=5820.22
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=4K --rw=randrw --rwmixread=75 --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=40g
    mybs: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    Jobs: 4 (f=4): [m(4)][20.1%][r=776MiB/s,w=259MiB/s][r=199k,w=66.2k IOPS][eta 02m:07s]
    Jobs: 4 (f=4): [m(4)][39.9%][r=785MiB/s,w=260MiB/s][r=201k,w=66.5k IOPS][eta 01m:35s]
    Jobs: 4 (f=4): [m(4)][59.9%][r=803MiB/s,w=270MiB/s][r=205k,w=69.1k IOPS][eta 01m:03s]
    Jobs: 4 (f=4): [m(4)][80.1%][r=808MiB/s,w=270MiB/s][r=207k,w=69.1k IOPS][eta 00m:31s]
    Jobs: 4 (f=4): [m(4)][99.4%][r=812MiB/s,w=273MiB/s][r=208k,w=69.9k IOPS][eta 00m:01s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21072: Sat Feb 16 10:37:08 2019
       read: IOPS=202k, BW=790MiB/s (828MB/s)(120GiB/155604msec)
        slat (nsec): min=1617, max=25132k, avg=11651.19, stdev=19195.76
        clat (nsec): min=484, max=31301k, avg=524183.11, stdev=545004.32
    
    
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=128K --rw=randrw --rwmixread=75 --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=40g
    mybs: (g=0): rw=randrw, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    Jobs: 4 (f=4): [m(4)][32.7%][r=1316MiB/s,w=418MiB/s][r=10.5k,w=3343 IOPS][eta 01m:06s]
    Jobs: 4 (f=4): [m(4)][65.6%][r=1318MiB/s,w=436MiB/s][r=10.5k,w=3489 IOPS][eta 00m:33s]
    Jobs: 4 (f=4): [m(4)][98.9%][r=1358MiB/s,w=438MiB/s][r=10.9k,w=3503 IOPS][eta 00m:01s]
    Jobs: 1 (f=0): [f(1),_(3)][100.0%][r=1334MiB/s,w=447MiB/s][r=10.7k,w=3575 IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21131: Sat Feb 16 10:40:00 2019
       read: IOPS=10.4k, BW=1296MiB/s (1359MB/s)(120GiB/94858msec)
        slat (usec): min=8, max=27157, avg=37.16, stdev=117.13
        clat (usec): min=182, max=76508, avg=9748.98, stdev=6240.33
         lat (usec): min=231, max=76526, avg=9786.65, stdev=6240.66
    
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=128K --rw=randread --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=80g
    mybs: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    mybs: Laying out IO file (1 file / 81920MiB)
    Jobs: 4 (f=4): [r(4)][19.3%][r=2567MiB/s,w=0KiB/s][r=20.5k,w=0 IOPS][eta 02m:14s]
    Jobs: 4 (f=4): [r(4)][43.4%][r=2547MiB/s,w=0KiB/s][r=20.4k,w=0 IOPS][eta 01m:22s]
    Jobs: 4 (f=4): [r(4)][67.6%][r=2551MiB/s,w=0KiB/s][r=20.4k,w=0 IOPS][eta 00m:45s]
    Jobs: 4 (f=4): [r(4)][91.9%][r=2555MiB/s,w=0KiB/s][r=20.4k,w=0 IOPS][eta 00m:11s]
    Jobs: 4 (f=4): [r(4)][100.0%][r=2554MiB/s,w=0KiB/s][r=20.4k,w=0 IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21251: Sat Feb 16 10:51:05 2019
       read: IOPS=19.4k, BW=2423MiB/s (2541MB/s)(320GiB/135240msec)
        slat (usec): min=7, max=3253, avg=26.96, stdev= 8.04
        clat (usec): min=219, max=22917, avg=6236.45, stdev=3332.59
    
    
    root@ubuntu:/media/ubuntu/p3605# fio --filename=out.fiod --ioengine=libaio --direct=1 --norandommap --randrepeat=0 --runtime=300 --blocksize=128K --rw=randwrite --iodepth=32 --numjobs=4 --group_reporting --eta-newline=30s --name=mybs --size=80g
    mybs: (g=0): rw=randwrite, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=libaio, iodepth=32
    ...
    fio-3.1
    Starting 4 processes
    Jobs: 4 (f=4): [w(4)][15.5%][r=0KiB/s,w=1624MiB/s][r=0,w=12.0k IOPS][eta 02m:54s]
    Jobs: 4 (f=4): [w(4)][30.6%][r=0KiB/s,w=1599MiB/s][r=0,w=12.8k IOPS][eta 02m:23s]
    Jobs: 4 (f=4): [w(4)][45.6%][r=0KiB/s,w=1580MiB/s][r=0,w=12.6k IOPS][eta 01m:52s]
    Jobs: 4 (f=4): [w(4)][60.7%][r=0KiB/s,w=1601MiB/s][r=0,w=12.8k IOPS][eta 01m:21s]
    Jobs: 4 (f=4): [w(4)][75.7%][r=0KiB/s,w=1600MiB/s][r=0,w=12.8k IOPS][eta 00m:50s]
    Jobs: 4 (f=4): [w(4)][90.8%][r=0KiB/s,w=1596MiB/s][r=0,w=12.8k IOPS][eta 00m:19s]
    Jobs: 4 (f=4): [w(4)][100.0%][r=0KiB/s,w=1594MiB/s][r=0,w=12.8k IOPS][eta 00m:00s]
    mybs: (groupid=0, jobs=4): err= 0: pid=21263: Sat Feb 16 10:55:26 2019
      write: IOPS=12.7k, BW=1587MiB/s (1664MB/s)(320GiB/206495msec)
        slat (usec): min=9, max=14945, avg=36.60, stdev=63.32
        clat (usec): min=2, max=37195, avg=10040.04, stdev=5912.15
         lat (usec): min=61, max=37222, avg=10077.11, stdev=5911.92
    
    
    
    
    
     
    #13
    Last edited: Feb 16, 2019
  14. metril

    metril New Member

    Joined:
    Oct 1, 2015
    Messages:
    14
    Likes Received:
    3
    Good. I bought 23x from the seller at $135 each.
     
    #14
    james23 likes this.
  15. james23

    james23 Active Member

    Joined:
    Nov 18, 2014
    Messages:
    372
    Likes Received:
    56
    so wow, we are now ~ 0.085 $ per gb for enterprise nvme (and pretty high WE nand). amazing price! that’s going to be a killer setup.

    if you don’t mind, how do you plan on using them? san/nas (which os / disk layout?), or are they going to be used individually?

    thanks
     
    #15
  16. metril

    metril New Member

    Joined:
    Oct 1, 2015
    Messages:
    14
    Likes Received:
    3
    They going into a supermicro NVMe storage server to serve me some fast storage for fun and games. I have not decided on the OS or the storage layout yet.
     
    #16
  17. Patrick

    Patrick Administrator
    Staff Member

    Joined:
    Dec 21, 2010
    Messages:
    11,384
    Likes Received:
    4,320
    I tried 8x at that price. Seller countered at $225 per
     
    #17
  18. K D

    K D Well-Known Member

    Joined:
    Dec 24, 2016
    Messages:
    1,397
    Likes Received:
    293
    Me too. Couldn't get him to go lower than 225.
     
    #18
  19. mimino

    mimino Member

    Joined:
    Nov 2, 2018
    Messages:
    31
    Likes Received:
    4
    Looks like a group buy would be a good solution here. I tried to get 4x250 and got turned down.
     
    #19
  20. Marsh

    Marsh Moderator

    Joined:
    May 12, 2013
    Messages:
    1,989
    Likes Received:
    912
    Is there a pattern ?

    Offer accepted
    2 x $200
    4 x $200
    23 x $135

    Offer rejected
    8 x $135
    4 x $250

    Counter-offer
    8 x $225
     
    #20
Similar Threads: 16TB intel
Forum Title Date
Great Deals Enterprise ssd up to 16TB Jan 29, 2019
Great Deals Amazon warehouse deals pricing error? Buffalo terastation 5410 incl 16tb Nov 20, 2017
Great Deals WD 16TB My Book Duo Desktop RAID External Hard Drive USB 3.0 $499.99 Nov 30, 2016
Great Deals WD My Book Duo 16TB 500€ Aug 23, 2016
Great Deals Intel 660p SSD 1TB M2 SSD (QLC) Friday at 8:22 AM

Share This Page