ML
    • Recent
    • Categories
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Cannot decide between 1U servers for growing company

    IT Discussion
    18
    246
    134.9k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • scottalanmillerS
      scottalanmiller @ntoxicator
      last edited by

      @ntoxicator said:

      I've moved VM's on Citrix Xen Server and Storage to another LUN at the time (when installed the 2nd Synology)

      It saturated the network.

      The current SuperMicro 1U server only has 2 Intel NIC cards. I have them bonded via Xen Center and LACP enabled.

      Yeah, storage migrations will do that 🙂 That's why you want block storage on a dedicated SAN if possible so that it uses its own "back channel" whenever possible so that it doesn't impact other things.

      1 Reply Last reply Reply Quote 1
      • scottalanmillerS
        scottalanmiller @ntoxicator
        last edited by

        @ntoxicator said:

        Also my goal was to migrate to NFS storage away from iSCSI

        as dealing with the RAW image or .cow2 image file is hell of alot easier.

        Agreed and good plan 🙂

        1 Reply Last reply Reply Quote 0
        • ntoxicatorN
          ntoxicator
          last edited by

          I had the Synology NAS setup as Block Level storage for the Volume that serves out the ISCSI Luns. mehhhhh.

          the complications! Lol.

          This is why I was wanting to move to all new design and setup being that I already essentially have data on a centralized setup.

          Could I get away with dual Synology 12-bay NAS units? (running in HA/replication).Probably

          I was thinking about having 10Gbe backbone/interconnect for the NAS + The VM Node servers. So all that traffic rides on the 10GBE backbone and would not touch the 1Gbe switches.

          AconboyA scottalanmillerS 2 Replies Last reply Reply Quote 2
          • scottalanmillerS
            scottalanmiller @ntoxicator
            last edited by

            @ntoxicator said:

            @Aconboy said:

            HC3

            How long has HC3 scale been avail? Today is my first time hearing and being introduced with the option

            Quite some time, they aren't new. They were a storage vendor before moving into hyperconvergence but they are one of the leaders in the HC space. They've been around longer than the terminology 🙂 HC is just starting to hit its stride in the market, though, so you'll start hearing about Scale and their competitors more and more in the near future.

            1 Reply Last reply Reply Quote 0
            • scottalanmillerS
              scottalanmiller @dafyre
              last edited by

              @dafyre said:

              The Scale systems are excellent. I know NTG has one. I've worked with their systems a couple of years ago, and the performance was night & day VS VMware and a similarly sized SAN. And their systems work really well.

              Scale especially kicks butt for Windows performance because of their stack.

              travisdh1T 1 Reply Last reply Reply Quote 0
              • AconboyA
                Aconboy @ntoxicator
                last edited by

                @ntoxicator - yup, we do that too - 10gig out of band from the LAN path for storage stack handling and cluster self-awareness.

                scottalanmillerS 1 Reply Last reply Reply Quote 1
                • scottalanmillerS
                  scottalanmiller @ntoxicator
                  last edited by

                  @ntoxicator said:

                  Could I get away with dual Synology 12-bay NAS units? (running in HA/replication).

                  Well here is the issue there.... If you are doing this as a traditional NAS (NFS or SMB) for file shares like mapped Windows drives or automounted home directories for Linux users this works reliably and beautifully. The failover is smooth and transparent.

                  From shops that have tested this with virtualization, the failover is not fast enough and the VMs typically fail so that you don't get the failover that you are hoping for but instead an outage with potential of corruption. The failover is just not fast enough (at least in real world tested scenarios) to use with VM HA.

                  1 Reply Last reply Reply Quote 0
                  • scottalanmillerS
                    scottalanmiller @Aconboy
                    last edited by

                    @Aconboy said:

                    @ntoxicator - yup, we do that too - 10gig out of band from the LAN path for storage stack handling and cluster self-awareness.

                    Yeah, we have a 10GigE fiber switching stack just to handle our OOB Scale communications! Talk about throughput capabilities!

                    1 Reply Last reply Reply Quote 0
                    • ntoxicatorN
                      ntoxicator
                      last edited by

                      So if anyone can explain to me

                      To do away with centralized storage such as what we have now and I've been moving to. I suppose this is what I've grown use to.

                      In order to have localized storage at the node/hypervisor level. one or many of the hypervisors would be storing all the data and sharing out the NFS? Then its replicated between? probably with DRBD Storage.

                      however, would would this be done with Citrix Xen Server for instance?

                      AconboyA coliverC scottalanmillerS DustinB3403D 4 Replies Last reply Reply Quote 0
                      • DashrenderD
                        Dashrender @ntoxicator
                        last edited by

                        @ntoxicator said:

                        @scottalanmiller

                        Essentially What I was looking to do was KVM / VM with complete HA.

                        I'm uncertain about keeping data local to individual servers. maybe because I have no experience with localized storage in an HA environment? Its all been shared centralized storage.

                        But you don't HA today. At least not at the storage level.

                        1 Reply Last reply Reply Quote 1
                        • DashrenderD
                          Dashrender @ntoxicator
                          last edited by

                          @ntoxicator said:

                          shit me for getting torn to shreds on here. Pissing contest.

                          Its much easier to verbalize than type out exact specifics.

                          What i meant by "I just seen as Windows iSCSI initiator working much better; more manageable and not limited."

                          Am I currently using windows iSCSI initator? NO
                          Do I wish I was using it: Yes?

                          Why: Because I feel it would be easier to manage and connect an iSCSI LUN as localized storage and data storage. The larger 2TB storage holds all the windows network shares and user profile data.... thats the problem.

                          So disconnect the LUN from the XenServer and connect it directly to ProxMox, then give that drive to the windows VM. does that not work in ProxMox?

                          scottalanmillerS 1 Reply Last reply Reply Quote 0
                          • AconboyA
                            Aconboy @ntoxicator
                            last edited by

                            @ntoxicator said:

                            So if anyone can explain to me

                            To do away with centralized storage such as what we have now and I've been moving to. I suppose this is what I've grown use to.

                            In order to have localized storage at the node/hypervisor level. one or many of the hypervisors would be storing all the data and sharing out the NFS? Then its replicated between? probably with DRBD Storage.

                            however, would would this be done with Citrix Xen Server for instance?

                            there are a couple of methods used there in the hyperconverged space. Two different schools of thought emerged on how best to simplify the architecture while maintaining the benefits of virtualization.

                            1. Simply virtualize the SAN and it's controllers - also known as pulling the SAN into the servers. The VSA or Virtual San Appliance approach was developed to move the SAN up into the host servers through the use of a virtual machine. This did in fact simplify things like implementation and management by eliminating the separate SAN. However, it didn't do much to simplify the data path or regain efficiency. The VSA consumed significant host resources (CPU and RAM), still used storage protocols, and complicated the path to disk by turning the IO path from application->RAM->Disk into application->RAM->hypervisor->RAM->SAN controller VM->RAM->hypervisor->RAM->write-cache SSD->Disk.

                            2. Eliminate the dedicated servers, storage protocol overhead, resources consumed and associated gear by moving the hypervisor directly into the OS of the storage platform as a set of kernel modules, thereby simplifying the architecture dramatically while regaining the efficiency originally promised by Virtualization.

                            1 Reply Last reply Reply Quote 0
                            • coliverC
                              coliver @ntoxicator
                              last edited by

                              @ntoxicator said:

                              So if anyone can explain to me

                              To do away with centralized storage such as what we have now and I've been moving to. I suppose this is what I've grown use to.

                              In order to have localized storage at the node/hypervisor level. one or many of the hypervisors would be storing all the data and sharing out the NFS? Then its replicated between? probably with DRBD Storage.

                              however, would would this be done with Citrix Xen Server for instance?

                              You could also look at Starwinds Virtual SAN. Which could do this as well.

                              scottalanmillerS 1 Reply Last reply Reply Quote 1
                              • DashrenderD
                                Dashrender @scottalanmiller
                                last edited by

                                @scottalanmiller said:

                                It is that it is on a LUN now that is limiting you. If it was on a NAS instead of a SAN, you'd have more options.

                                Scott, where's your link explaining the difference?

                                coliverC scottalanmillerS 2 Replies Last reply Reply Quote 0
                                • coliverC
                                  coliver @Dashrender
                                  last edited by coliver

                                  @Dashrender said:

                                  @scottalanmiller said:

                                  It is that it is on a LUN now that is limiting you. If it was on a NAS instead of a SAN, you'd have more options.

                                  Scott, where's your link explaining the difference?

                                  One is block the other is file?

                                  DashrenderD 1 Reply Last reply Reply Quote 0
                                  • DashrenderD
                                    Dashrender @coliver
                                    last edited by

                                    @coliver said:

                                    @Dashrender said:

                                    @scottalanmiller said:

                                    It is that it is on a LUN now that is limiting you. If it was on a NAS instead of a SAN, you'd have more options.

                                    Scott, where's your link explaining the difference?

                                    One is block the other is file?

                                    The purpose of my post was to show that the OP was using his NAS as both NAS and SAN simultaneously to ensure Scott's point wasn't being lost when he was indicating the use of a SAN or NAS.

                                    1 Reply Last reply Reply Quote 1
                                    • scottalanmillerS
                                      scottalanmiller @ntoxicator
                                      last edited by

                                      @ntoxicator said:

                                      So if anyone can explain to me

                                      To do away with centralized storage such as what we have now and I've been moving to. I suppose this is what I've grown use to.

                                      In order to have localized storage at the node/hypervisor level. one or many of the hypervisors would be storing all the data and sharing out the NFS? Then its replicated between? probably with DRBD Storage.

                                      however, would would this be done with Citrix Xen Server for instance?

                                      No NFS, SMB, iSCSI or anything involved with the DRBD scenario. It's raw block storage replication all the way down in the stack. Never need to look at high level stuff. The storage is visible to both nodes at the same time locally. No sharing protocols at all.

                                      With XenServer you could do this yourself but for best results you'd likely use HA-Lizard which is a set of tools around DRBD on XenServer handling all of the complication for you. Not only is it free, but the HA-Lizard team participates here in the community a little.

                                      1 Reply Last reply Reply Quote 0
                                      • scottalanmillerS
                                        scottalanmiller @Dashrender
                                        last edited by

                                        @Dashrender said:

                                        @ntoxicator said:

                                        shit me for getting torn to shreds on here. Pissing contest.

                                        Its much easier to verbalize than type out exact specifics.

                                        What i meant by "I just seen as Windows iSCSI initiator working much better; more manageable and not limited."

                                        Am I currently using windows iSCSI initator? NO
                                        Do I wish I was using it: Yes?

                                        Why: Because I feel it would be easier to manage and connect an iSCSI LUN as localized storage and data storage. The larger 2TB storage holds all the windows network shares and user profile data.... thats the problem.

                                        So disconnect the LUN from the XenServer and connect it directly to ProxMox, then give that drive to the windows VM. does that not work in ProxMox?

                                        SANs do not work that way.

                                        DashrenderD 1 Reply Last reply Reply Quote 0
                                        • scottalanmillerS
                                          scottalanmiller @coliver
                                          last edited by

                                          @coliver said:

                                          @ntoxicator said:

                                          So if anyone can explain to me

                                          To do away with centralized storage such as what we have now and I've been moving to. I suppose this is what I've grown use to.

                                          In order to have localized storage at the node/hypervisor level. one or many of the hypervisors would be storing all the data and sharing out the NFS? Then its replicated between? probably with DRBD Storage.

                                          however, would would this be done with Citrix Xen Server for instance?

                                          You could also look at Starwinds Virtual SAN. Which could do this as well.

                                          But is limited to Hyper-V for best results and can make due in the VMware world but is inferior (but you shouldn't be looking at ESXi anyway so not a big deal.)

                                          coliverC 1 Reply Last reply Reply Quote 0
                                          • DashrenderD
                                            Dashrender @scottalanmiller
                                            last edited by

                                            @scottalanmiller said:

                                            @Dashrender said:

                                            @ntoxicator said:

                                            shit me for getting torn to shreds on here. Pissing contest.

                                            Its much easier to verbalize than type out exact specifics.

                                            What i meant by "I just seen as Windows iSCSI initiator working much better; more manageable and not limited."

                                            Am I currently using windows iSCSI initator? NO
                                            Do I wish I was using it: Yes?

                                            Why: Because I feel it would be easier to manage and connect an iSCSI LUN as localized storage and data storage. The larger 2TB storage holds all the windows network shares and user profile data.... thats the problem.

                                            So disconnect the LUN from the XenServer and connect it directly to ProxMox, then give that drive to the windows VM. does that not work in ProxMox?

                                            SANs do not work that way.

                                            Do we need another thread for an explanation of why not? I don't understand why not, at least not as stated.

                                            scottalanmillerS 1 Reply Last reply Reply Quote 1
                                            • 1
                                            • 2
                                            • 3
                                            • 4
                                            • 5
                                            • 6
                                            • 12
                                            • 13
                                            • 4 / 13
                                            • First post
                                              Last post