ML
    • Recent
    • Categories
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Xen Server 6.5 + Xen Orchestra w. HA & SAN

    IT Discussion
    8
    112
    33.4k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • DashrenderD
      Dashrender @ntoxicator
      last edited by

      @ntoxicator said:

      @scottalanmiller said:

      Thanks scott. makes sense and I understand.

      So again, just have to spec out a 2U server (I assume 2U). With the required disk space which would hold us out for 5+ years. I am going to say we would well over 5TB+ to be safe.

      Could use 600GB or larger SAS drives with hardware raid controller. Or some enterprise level 7200RPM drives? I'm unsure how folks feel about those.

      Is there a specific reason you mention the height of a server? Do you have limitations in your cabinet that limit how tall your servers are?
      In my office I have 1 full height cabinet, 42 U worth of space. Could I have gotten away with a half cabinet, definitely, but I have what I have. I'm using 8 U for UPSs, 4U (two 2U servers) for 10 year old EHR, 4U (two 2U servers) for hypervisor hosts, 2U for a Drobo and 1U for a network switch and 1U for a KVM panel. Grand total 20U. I still have over half the rack left over for expansion.

      If I were looking at new servers, the height of the server would be the least of my concerns. Granted you can get 2U servers today that hold nearly 20 disks, it wouldn't matter to me if it was 4U because I have the space.

      The size of the drives you buy will be dependent upon a few factors. What do you need for IOPs? If you have low IOPs needs, why not buy 4 TB drives? 4 of them in RAID 10 would give you 8 TB of usable space. If you need higher IOPS, perhaps eight 2 TB drives in RAID 10 would be better, still leaving you with 8 TB usable space.

      To determine your IOPs requirement, you could get a Dell DPack run against your system. You just have to ignore the sales people trying to sell you a SAN and remember, Dell isn't trying to be your friend, they are trying to extract money from you. Ignore their SAN recommendation, and post the results to a place like ML to get help/suggestions on what to get.

      Another option would be to hire a firm to do all of this specing for you. They will run the tools and then recommend a system. This is a situation where you are paying someone for their opinion, preferably someone who isn't trying to sell you anything else. This way they understand that they are making money on their opinion/suggestion, not on the hopes of selling you hardware.

      1 Reply Last reply Reply Quote 2
      • scottalanmillerS
        scottalanmiller @ntoxicator
        last edited by

        @ntoxicator said:

        Could use 600GB or larger SAS drives with hardware raid controller. Or some enterprise level 7200RPM drives? I'm unsure how folks feel about those.

        The spindle speed is just spindle speed. If 7200 RPM drives have the IOPS that you need, they are just as good (or better since they are cheaper and more reliable.)

        1 Reply Last reply Reply Quote 1
        • ntoxicatorN
          ntoxicator
          last edited by

          Question

          HA-Lizard HA-iSCSI interface between the two(2) servers..

          bonded GigE link?

          or a single DA 10Gbe cable between the hosts? I suppose, bonded GigE is plenty sufficient due to disk write speeds? (non SSD)

          1 Reply Last reply Reply Quote 0
          • scottalanmillerS
            scottalanmiller
            last edited by

            Can't bond SAN links.

            1 Reply Last reply Reply Quote 0
            • ntoxicatorN
              ntoxicator
              last edited by

              I could be incorrect on the terminology or reference. I was reading one of their posted documents.

              This would be the DRBD interface / IP link between the two(2) nodes

              scottalanmillerS 1 Reply Last reply Reply Quote 0
              • ntoxicatorN
                ntoxicator
                last edited by

                http://www.halizard.com/images/pdf/iscsi-ha_2-node_cluster_howto_1.2.12_final.pdf

                1 Reply Last reply Reply Quote 0
                • scottalanmillerS
                  scottalanmiller @ntoxicator
                  last edited by

                  @ntoxicator said:

                  This would be the DRBD interface / IP link between the two(2) nodes

                  DRBD is the protocol there, no iSCSI. It's not SAN or anything like that.

                  1 Reply Last reply Reply Quote 0
                  • ntoxicatorN
                    ntoxicator
                    last edited by

                    Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                    So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                    DashrenderD scottalanmillerS 3 Replies Last reply Reply Quote 0
                    • DashrenderD
                      Dashrender @ntoxicator
                      last edited by

                      @ntoxicator said:

                      Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                      So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                      Well 10 Ge never hurts...

                      1 Reply Last reply Reply Quote 0
                      • scottalanmillerS
                        scottalanmiller @ntoxicator
                        last edited by

                        @ntoxicator said:

                        Ok. Their documentation says the DRBD interface to be bonded from within Xencenter (per documentation). Am I wrong here?

                        That seems fine. DRBD works differently than iSCSI. They are not related protocols.

                        1 Reply Last reply Reply Quote 0
                        • scottalanmillerS
                          scottalanmiller @ntoxicator
                          last edited by

                          @ntoxicator said:

                          So the ethernet link between the two nodes, im sure GigE is plenty enough bandwidth? or 10Gige not hurt?

                          What is the bandwidth of the storage? You will be limited to GigE throughput speeds, that's 1Gb/s, for writes. That's a fraction of what SATA and SAS can do.

                          1 Reply Last reply Reply Quote 0
                          • scottalanmillerS
                            scottalanmiller
                            last edited by

                            Are you sure that all of this makes sense in your environment? This is two orders of magnitude from where you have been in the past. It isn't normal to have an LA (low availability) environment and get by for a long time and suddenly leap to HA. Why not just go to standard availability? It's a full order of magnitude safer than where you have been in the past, almost zero effort (and no risk from that lack of effort... simple is your friend) and less than half the price of doing HA.

                            SA is the only clear win... tons safer, tons cheaper. HA is tons safer for sure, but costs more and doesn't make sense given what was deemed acceptable in the past.

                            ntoxicatorN 1 Reply Last reply Reply Quote 1
                            • ntoxicatorN
                              ntoxicator
                              last edited by

                              Looking to develop hardware costs and quotes for new equipment. Company wants to grow employee's to 500+ by year 2020. Need to have reliable servers hosting VM's

                              If the primary xenserver host fails.. then what? We have a day + of downtime waiting for server to come back online?

                              Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                              scottalanmillerS 3 Replies Last reply Reply Quote 0
                              • ntoxicatorN
                                ntoxicator @scottalanmiller
                                last edited by

                                @scottalanmiller said:

                                Are you sure that all of this makes sense in your environment? This is two orders of magnitude from where you have been in the past. It isn't normal to have an LA (low availability) environment and get by for a long time and suddenly leap to HA. Why not just go to standard availability? It's a full order of magnitude safer than where you have been in the past, almost zero effort (and no risk from that lack of effort... simple is your friend) and less than half the price of doing HA.

                                SA is the only clear win... tons safer, tons cheaper. HA is tons safer for sure, but costs more and doesn't make sense given what was deemed acceptable in the past.

                                also to mention, I've been hammering HA setup for awhile to management; for peace of mind and rest-easy at night. Yes, we've been getting along with low availability type setup for now. But as the resource usages increase; i feel the need for HA setup with dual nodes.

                                scottalanmillerS 1 Reply Last reply Reply Quote 0
                                • scottalanmillerS
                                  scottalanmiller @ntoxicator
                                  last edited by

                                  @ntoxicator said:

                                  Looking to develop hardware costs and quotes for new equipment. Company wants to grow employee's to 500+ by year 2020. Need to have reliable servers hosting VM's

                                  That's fine. But that doesn't suggest HA at all.

                                  1 Reply Last reply Reply Quote 1
                                  • scottalanmillerS
                                    scottalanmiller @ntoxicator
                                    last edited by

                                    @ntoxicator said:

                                    If the primary xenserver host fails.. then what? We have a day + of downtime waiting for server to come back online?

                                    This is not how you discuss risk. This tells me that HA is not needed. This isn't how a "we need HA" discussion would start.

                                    1 Reply Last reply Reply Quote 0
                                    • scottalanmillerS
                                      scottalanmiller @ntoxicator
                                      last edited by

                                      @ntoxicator said:

                                      Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                                      This strongly says that HA is a bad idea as it requires a lot more skill, documentation, knowledge, etc.

                                      scottalanmillerS 1 Reply Last reply Reply Quote 0
                                      • scottalanmillerS
                                        scottalanmiller @ntoxicator
                                        last edited by

                                        @ntoxicator said:

                                        But as the resource usages increase; i feel the need for HA setup with dual nodes.

                                        Resource utilization does not lead directly to needing higher availability. Wall St. firms don't need HA on billion dollar trading systems, why do you feel that moving to 500 users warrants in for you but hundreds of thousands of users does not warrant it for them?

                                        That's not to say that HA can't be right for you, I'm saying that you aren't thinking about it in the way that you should if you were going to determine that HA was needed.

                                        1 Reply Last reply Reply Quote 0
                                        • scottalanmillerS
                                          scottalanmiller @scottalanmiller
                                          last edited by

                                          @scottalanmiller said:

                                          @ntoxicator said:

                                          Looking to future proof so can be in production for next 5 years. I may not be here with company in next 5 years, so want to leave behind a good setup.

                                          This strongly says that HA is a bad idea as it requires a lot more skill, documentation, knowledge, etc.

                                          And if HA is determined to be needed, it needs to be an appliance, not a solution that they will be unable to support.

                                          1 Reply Last reply Reply Quote 0
                                          • scottalanmillerS
                                            scottalanmiller
                                            last edited by

                                            Server planning for five years out is relatively risky. Not bad, but we know just about nothing about the future. And "the company wants to grow" should not lead us to spending today. A 200 person company buying hardware like they are a 500 person company is the hallmark of a company that is never going to get to 500. Projections like that are not things that IT acts on, that would lead to some dangerous stuff.

                                            1 Reply Last reply Reply Quote 0
                                            • 1
                                            • 2
                                            • 3
                                            • 4
                                            • 5
                                            • 6
                                            • 1 / 6
                                            • First post
                                              Last post