Napp-It AIO on 6.5[6.7] and Windows 10 Pro

Discussion in 'Solaris, Nexenta, OpenIndiana, and napp-it' started by pc-tecky, Jul 5, 2018.

  1. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    @gea, Can you explain this behavior?

    I just got done installing ESXi 6.7 and Napp-It (via OVA, dated May 2018), then stepped down to ESXi 6.5U2 and restored/registered Napp-It with ESXi 6.5. I can get files from Windows 10 Pro to the file share on Napp-It, but I can't get any sized file (little documents and larger ISOs) from Napp-It back to Windows 10 with either a network Error 0x8007003B or simply hangs/times out.

    On another note, I think I'm hosed on my older FreeNAS setup, while I still have the data drives mostly untouched, I can't find the OS drive, and therefore can't restore the prior ZFS array.

    ----

    Got it working, 6 - 7 hours later.. Added SMBv1 back to Windows 10. Double checked my workgroup name. And lastly, made a change to reflect private network as private, and not public. I suspect the last change made all the difference. Buried and not as simple straight forward as I recall it being in Windows 7 and prior.

    Next question though is how to allow ESXi access to a portion of the ZFS data pool available to host some VMs?
     
    #1
    Last edited: Jul 6, 2018
  2. gea

    gea Well-Known Member

    Joined:
    Dec 31, 2010
    Messages:
    1,885
    Likes Received:
    626
    You can either share a ZFS filesystem via NFS (what I prefer) or via iSCSI.
    In ESXi you can use either as a datastore for VMs
     
    #2
  3. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    I've been trying to share the ZFS via NFS from within Napp-It. But so far I still only have the local SSD as only ESXi VM storage option. What do I need to do? Where do I go? I'm pretty sure I have NFS running from within Napp-It, whether secured or not is another issue..
     
    #3
  4. gea

    gea Well-Known Member

    Joined:
    Dec 31, 2010
    Messages:
    1,885
    Likes Received:
    626
    1.
    A napp-it network interface must be in the same ip range and vswitch than the ESXi management interface

    2.
    Enable NFS with a everyone@=modify ACL set

    3.
    add the NFS share in ESXi like 192.168.1.10:/poolname/filesystem

    To make it secure, add a firewall rule for NFS or use iSCSI restricted to an ip range. For best security use a dedicated napp-it for VM storage only (ex SSD pool) and use a dedicated vswitch for all management interfaces, no public access.

    For general filer use, add a second napp-it VM with mass-storage and a second nic to the management LAN ex for backups.
     
    #4
  5. K D

    K D Well-Known Member

    Joined:
    Dec 24, 2016
    Messages:
    1,373
    Likes Received:
    286
  6. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    @K D, thanks.. missing something.. a few things actually..

    1) How did you go from default vSwitch0 (etc.) to custom names?

    2) maybe my downfall, I have a singular ZFS pool set for SAMBA/CIFS initally, then enabled NFS. Still not seeing how it's made available to ESXi.

    Looking at adding some complexities like hosting untangle NG or pfSense with a virtual ESXi External switch and a virtual ESXi Internal switch, and maybe an Internal NFS only switch or VLAN.

    Modem WiFi Adapter Switch <-> ESXi External switch <-> untanlge <-> ESXi Internal switch <-> external Cisco Switch for other physical boxes, my main PC, etc.
     
    #6
    Last edited: Jul 7, 2018
  7. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    blah! purple death? borked? time to restart the install mess again?
     
    #7
  8. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    While the virtual switches are 9000 MTU, the Napp-It virtual NICs remain at 1500 MTU. How do I configure this to be 9000? How do I change the IP address. Familiar with windows and Linux - this is a bit different..
     
    #8
  9. gea

    gea Well-Known Member

    Joined:
    Dec 31, 2010
    Messages:
    1,885
    Likes Received:
    626
    Basically you must disable the link, change MTU and re-enable.
    But as internal transfers within ESXi is pure software, MTU does not matter.

    I asume that you manage via the e1000 link and use the second vmxnet3 link for data transfers. In menu System > Network ETH you can then disable the link and set MTU

    From console commands, see
    napp-it // webbased ZFS NAS/SAN appliance for OmniOS, OpenIndiana, Solaris and Linux : OmniOS
    Setting Datalink Properties - Oracle Solaris Administration: Network Interfaces and Network Virtualization
     
    #9
  10. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    well that was a bit of a headache, but got the MTU set after disabling, setting MTU=9000, and then re-enable the interface. Then I lost IP access.

    ESXi VMKernel NICs
    vmk0 --> (10.1.1.4) Management Network
    vmk1 --> (10.1.100.251) Internal Storage Kernel
    vmk2 --> (10.1.100.105) Storage Network (if used, disassociates Napp-It 3rd NIC from Storage switch)

    Virtual Switches
    vSwitch0 (can't move, rename, or edit much??, MTU: 9000 ; VLAN ID: 0 )
    Storage (the internal network for storage, MTU: 9000 ; VLAN ID: 100)
    Internal ( ??, MTU: 9000 ; VLAN ID: 0 ; future use with pfSense or Untangle router OS, 2-ports in, 4-6 ports out)
    External ( ??, MTU: 9000 ; VLAN ID: 0 ; future use with pfSense or Untangle router OS, 2-ports in, 4-6 ports out)

    Port Groups
    VM Network --> vSwitch0 (default)
    Management Network --> vSwitch0 (default)
    Internal Storage Network --> vSwitch0 (tried following K D's writeup, doesn't click and I suspect some missing steps, common for them)
    Storage Network --> Storage

    OK, so how does the VMK1 with PortGroup 'Internal Storage Network' on vSwitch0 jump over to PortGroup 'Storage Network' on 'Storage' switch?? I suspect that it doesn't work quite that way.

    I'll have to start physically assigning IPs despite how much I enjoy the simplicity of how DHCP works and not having to worry about assigning IPs in the past.

    ----
    Alright, made some progress. Started Napp-It after having removed and then re-added VMK2 to the 'Storage' switch and removed VMK1 - needed to make the virtual physical connection all within the same realm. For whatever reason, the other way around kicked Napp-It off the storage switch. But I have had a few purple deaths as well.
     
    #10
  11. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    @gea , @K D - ok, so trying this instead and hoping maybe it works:

    ESXi VMKernel NICs
    vmk0 --> (10.1.1.4) Management Network
    vmk1 --> (10.1.100.105) Internal Storage Kernel
    ** got to watch this. why does the ESXi VMK1 NIC connection disassociates Napp-It's NIC connection to the Storage switch, or stranger yet, make the Storage switch disappear??

    Virtual Switches
    vSwitch0 (can't move, rename, or edit much??, MTU: 9000 ; VLAN ID: 0 )
    Storage (the internal network for storage, MTU: 9000 ; VLAN ID: 100)
    Internal ( ??, MTU: 9000 ; VLAN ID: 0 ; future use with pfSense or Untangle router OS, 2-ports in, 4-6 ports out)
    External ( ??, MTU: 9000 ; VLAN ID: 0 ; future use with pfSense or Untangle router OS, 2-ports in, 4-6 ports out)

    Port Groups
    VM Network --> vSwitch0 (default)
    Management Network --> vSwitch0 (default)
    Kernel Storage --> Storage (ESXi for NFS)
    Storage Network --> Storage (Napp-It NFS)

    root@napp-it-026:~#
    root@napp-it-026:~# ipadm show-addr
    ADDROBJ ~> TYPE ~> STATE ~> ADDR
    lo0/v4 ~> static ~> ok ~> 127.0.0.1/8
    lo0/v6 ~> static -> ok~> ::1/128
    e1000g1/v4 ~> dhcp ~> disabled ~> ?
    wmxNET3s0/v4 ~> dhcp ~> disabled ~> ?
    e1000g0/v4 ~> dhcp ~> disabled ~> ?
    e1000g2/v4 ~> static ~> disabled ~> 10.1.100.151/24
    root@napp-it-026:~# ipadm show-if
    IFNAME ~> STATE -> CURRENT -> PERSISTENT
    lo ~> ok ~> -m-v------46 ~> ---
    e1000g0 ~> down ~> bm------46 ~> -46
    e1000g1 ~> disabled ~> ---------- ~> -46
    e1000g2 ~> disabled ~> ---------- ~> -46
    root@napp-it-026:~# dladm show-link
    LINK ~> CLASS ~> MTU ~> STATE ~> BRIDGE ~> OVER
    e1000g0 ~> phys ~> 9000 ~> up ~> -- ~> --
    e1000g3 ~> phys ~> 1500 ~> unknown ~> -- ~> --
    e1000g4 ~> phys ~> 1500 ~> unknown ~> -- ~> --
    root@napp-it-026:~# dladm show-phys


    I don't think I have e1000g4, e1000g3, or vmxnet3s0 any longer, but the system commands show I still have these phantom handware still installed.. so why are the showing up? how do I remove them?

    ipadm disable-if -t *interface*
    dladm set-linkprop -p mtu=9000 *interface*
    ipadm delete-if *interface*
    ipadm enable-if *interface*
    ipadm show-if
     
    #11
  12. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    @gea , @K D , please explain the menu options (clear as mud for me atm)
    New datastore \\ Provide NFS mount details

    Name: ESXi NFS (use whatever name I please that will be meaningful?)
    NFS Server: 10.1.100.151 (the assigned static IP using VLAN ID: 100)
    NFS Share: (this would be?? a) a local mount point -or- b) a remote mount point)
    NFS Version: [X] NFS3 -or- [ ] NFS4
     
    #12
  13. K D

    K D Well-Known Member

    Joined:
    Dec 24, 2016
    Messages:
    1,373
    Likes Received:
    286
    [​IMG]
     
    #13
  14. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    my first casualty, a drive is dead..

    -----

    I do all of that and nothing so far..

    I go to ZFS Pool, and under the NFS colum of the share click to turn it 'ON'.. then where to for permissions, or is that it? if Napp-It is on 10.1.1.5, then how does 10.1.100.151 on second NIC for Napp-It work - is this a case of a dual 'homed' server?
     
    #14
    Last edited: Jul 9, 2018
  15. K D

    K D Well-Known Member

    Joined:
    Dec 24, 2016
    Messages:
    1,373
    Likes Received:
    286
    Are you able to ping the esxi host from the VM and vice versa?
     
    #15
  16. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    eurika! forward slash vs back slash? I don't know now which it was, but that was the hang up and I got it working... err!!!!! :mad::confused::confused:o_O:rolleyes: the entire mess with VLAN IDs, as far as I could tell was extra work that didn't work, couldn't ping between the static IP addresses at all, no connections, nothing really..
     
    #16
    Last edited: Jul 10, 2018
  17. pc-tecky

    pc-tecky Member

    Joined:
    May 1, 2013
    Messages:
    194
    Likes Received:
    22
    bringing this back to life... have an error with Napp-IT stating the ZFS pool can't be found and bascially no bootloader.. quick searches point to a shared drive cache sort of bug with an easy fix from mid-2017 as I recall.. now just need to restore this if I can..
     
    #17
Similar Threads: Napp-It 65[67]
Forum Title Date
Solaris, Nexenta, OpenIndiana, and napp-it Looking to update OmniOS/NAPP-IT from r151014 Oct 23, 2018
Solaris, Nexenta, OpenIndiana, and napp-it OmniOS/napp-it self-sign certificate Oct 4, 2018
Solaris, Nexenta, OpenIndiana, and napp-it Solaris (OmniOS) w/ Napp-It ZPool Share Permissions for CIFS [Solved] Sep 17, 2018
Solaris, Nexenta, OpenIndiana, and napp-it NAPP-IT bug: zpool add disks to mirror, pool incorrectly reported non-existant Aug 31, 2018
Solaris, Nexenta, OpenIndiana, and napp-it Windows Offline-Files and Napp-it SMB Aug 29, 2018

Share This Page