ML
    • Recent
    • Categories
    • Tags
    • Popular
    • Users
    • Groups
    • Register
    • Login

    Exablox testing journal

    IT Discussion
    7
    75
    27.6k
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as topic
    Log in to reply
    This topic has been deleted. Only users with topic management privileges can see it.
    • Reid CooperR
      Reid Cooper
      last edited by

      How many total "nodes" do you have? Two total, one in each ring?

      MattSpellerM 1 Reply Last reply Reply Quote 0
      • MattSpellerM
        MattSpeller @Reid Cooper
        last edited by

        @Reid-Cooper Aye, just 4 drives each - if they work out we have a buttload of video to dump on them which may mean more drives. There's a catch to that and I'll explain more later.

        Correct, just two "oneblox"

        1 Reply Last reply Reply Quote 0
        • MattSpellerM
          MattSpeller
          last edited by MattSpeller

          Initial impressions;

          • nice build quality
          • half rack units
          • they stack nicely with little feet
          • rackmounts works a treat but fit in our racks was pretty tight, had to do a bit of jiggling (our racks suck, YMMV)
          • drive bay doors are plastic, which should be fine if you're careful ($0.02 - for the $$$ some metal would have been nice, and more durable; we expect these to last 5 years and have multiple drive upgrades over time)
          • drive lights do not indicate drive activity, only a populated bay (possibly also failed drive)
          • came with all hardware required for rack or stack (unsure if this is standard)
          • very nicely packaged, I would expect these to survive severe shipping
          • LCD screen is useful (see pic on next post)
          1 Reply Last reply Reply Quote 2
          • MattSpellerM
            MattSpeller
            last edited by MattSpeller

            IMG_20150130_133340.jpg

            1 Reply Last reply Reply Quote 0
            • scottalanmillerS
              scottalanmiller
              last edited by

              I think of Exablox a little bit like Drobo for grown ups. Drobo is neat but really small scale only both in terms of total possible size (it tops out at twelve drives and that's the super expensive model), is extremely limited in power (very slow) and lacks the reliability features that you often want for critical storage. It has good niches, but it doesn't scale up at all.

              Exablox takes the same "super simple, just plug and go" approach but gives you massive scale and big time redundancy options so that you can use it for scenarios that need speed, scale and/or reliability. But there is still that ease of use that is similar to what Drobo was attempting to accomplish.

              1 Reply Last reply Reply Quote 1
              • MattSpellerM
                MattSpeller
                last edited by

                So, ran into a hiccup yesterday just before I left. Received an email from Exablox pointing to this link - one for each of the Oneblox units we have. 90 minutes later I received an email for one unit, that it had "cleared" the error (link was identical, no additional info provided). During this time I noticed no difference in it's performance or behaviour - watching it sync across the "ring" to the other unit I know it was still transferring data (of some kind). Currently I still have one unit displaying the error, the other has been cleared and it has not returned (~20 hours and counting).

                Picture of the error that was displayed on the LCD screen
                IMG_20150129_170626.jpg

                1 Reply Last reply Reply Quote 0
                • scottalanmillerS
                  scottalanmiller
                  last edited by

                  I like that LCD screen. That is cool.

                  MattSpellerM 1 Reply Last reply Reply Quote 0
                  • MattSpellerM
                    MattSpeller @scottalanmiller
                    last edited by

                    @scottalanmiller Me too - it's got just enough room for important info

                    1 Reply Last reply Reply Quote 2
                    • MattSpellerM
                      MattSpeller
                      last edited by

                      When I received the error I was attempting to simulate a common experience - big data dump onto the file share. To simulate this I have the "local" oneblox unit connected into our 100mbit lan properly, the other unit is connected first to a 10mbit hub to simulate the speed and quality of our connection between sites. Also on that hub is a laptop with Wireshark so I can learn more about networking and watch to see what these things do and why they're so chatty - thats next up.

                      scottalanmillerS 1 Reply Last reply Reply Quote 2
                      • MattSpellerM
                        MattSpeller
                        last edited by MattSpeller

                        Chatty little boxes! These things call home quite a bit. It's not a huge quantity of traffic (around 300mb/day/oneblox = 600mb/day total for me). Keep in mind that is without anyone actually accessing them or doing anything to them. I'm not sure if that will effect their level of communication or not. This screen cap is over 24h period.

                        Capture.JPG

                        SeanExabloxS 2 Replies Last reply Reply Quote 0
                        • scottalanmillerS
                          scottalanmiller
                          last edited by

                          My guess, but it is only a guess, is that that number is diagnostics and will grow slightly but be generally pretty consistent over time. Maybe @SeanExablox can shed some light on that.

                          SeanExabloxS 2 Replies Last reply Reply Quote 0
                          • scottalanmillerS
                            scottalanmiller @MattSpeller
                            last edited by

                            @MattSpeller said:

                            When I received the error I was attempting to simulate a common experience - big data dump onto the file share. To simulate this I have the "local" oneblox unit connected into our 100mbit lan properly, the other unit is connected first to a 10mbit hub to simulate the speed and quality of our connection between sites. Also on that hub is a laptop with Wireshark so I can learn more about networking and watch to see what these things do and why they're so chatty - thats next up.

                            Sounds like some serious testing. I like where this is all going.

                            MattSpellerM 1 Reply Last reply Reply Quote 0
                            • SeanExabloxS
                              SeanExablox @scottalanmiller
                              last edited by

                              This post is deleted!
                              1 Reply Last reply Reply Quote 0
                              • SeanExabloxS
                                SeanExablox @MattSpeller
                                last edited by

                                This post is deleted!
                                1 Reply Last reply Reply Quote 1
                                • SeanExabloxS
                                  SeanExablox @scottalanmiller
                                  last edited by

                                  @scottalanmiller

                                  @scottalanmiller said:

                                  My guess, but it is only a guess, is that that number is diagnostics and will grow slightly but be generally pretty consistent over time. Maybe @SeanExablox can shed some light on that.

                                  @scottalanmiller you're correct. Because our management is cloud-based the heartbeat and metadata comprise the payload sent over a 24 hour period. The larger and more complex the environment (#shares, users, events, etc) the amount traffic OneBlox and OneSystem pass back and forth will increase. Having said that it should be a very modest increase in traffic and it's fairly evenly distributed throughout the day.

                                  1 Reply Last reply Reply Quote 3
                                  • SeanExabloxS
                                    SeanExablox @MattSpeller
                                    last edited by

                                    @MattSpeller

                                    @MattSpeller said:

                                    Chatty little boxes! These things call home quite a bit. It's not a huge quantity of traffic (around 300mb/day/oneblox = 600mb/day total for me). Keep in mind that is without anyone actually accessing them or doing anything to them. I'm not sure if that will effect their level of communication or not. This screen cap is over 24h period.

                                    Capture.JPG

                                    @MattSpeller user access and storage capacity won't really impact the amount of data that OneBlox and OneSystem send to each other. If you do see a significant uptick, please let me know.

                                    PS. apologies for the delete/repost, neglected to quote the original post.

                                    1 Reply Last reply Reply Quote 3
                                    • MattSpellerM
                                      MattSpeller @scottalanmiller
                                      last edited by

                                      @scottalanmiller said:

                                      Sounds like some serious testing. I like where this is all going.

                                      I'd appreciate suggestions for other things to try!! I have a ton to learn about networking and this is a great opportunity. I will post about the hardware angle as well.

                                      1 Reply Last reply Reply Quote 0
                                      • coliverC
                                        coliver
                                        last edited by

                                        I will be interested to see what you learn. This looks like a cool technology.

                                        1 Reply Last reply Reply Quote 2
                                        • scottalanmillerS
                                          scottalanmiller
                                          last edited by

                                          I am really interested to see this running production virtualization workloads via NFS.

                                          MattSpellerM SeanExabloxS 2 Replies Last reply Reply Quote 2
                                          • MattSpellerM
                                            MattSpeller
                                            last edited by

                                            Ignorance ahoy! I really didn't understand at all how these things work and the following will probably illuminate that, hopefully I'll learn more!

                                            My primary concern (curiosity? frustration?) is how the bleeping heck do they claim 100% space utilization. So this means you toss in a 1TB drive, it appears with 1TB more free space for you to use. This is basically anathema to me when they claim to have your data backed up and safe - to the point where you can just yank a drive out and it'll continue on it's merry way. As we all know, with RAID to get your data backed up it's usually a 50% space penalty (RAID1 / 10). Knowing that the answer probably isn't "witchcraft" I set about trying to figure it all out. The following is my super high level guess and assumptions and I'm probably wrong about half of it.

                                            Object Based Storage is the heart of the answer, I'm pretty sure. My one sentence understand of it is: cut data into chunks, number the chunks, hash them?, store them across a couple (3?) drives. It appears that this is very similar to RAID in that the data is duped; where I know I'm missing something key is that you don't get 1:1 utilization if you're duping data anywhere. Impossible right? Yes. My best guess is that they do take a penalty from duping data, but they MUST recoup it somewhere!

                                            Taking the units I have as the working example: put in 4x4TB drives, got 14.4TB usable storage out. That's not unreasonable to lose some capacity to file system, formatting, whatever. We do know that they use de-dupe and this is primarily how I suppose they recoup the space and claim 1:1. I have checked on the management page and our units are claiming a de-dupe ratio of 1.3. According to "My Computer" I have 2.23TB used currently. Now, here is where the slight of hand comes in - I have (I hope... I didn't actually keep careful track) actually put 2.23TB of data on the thing. AH HAH! Witchcraft this is not! It's just not displaying the space it gained by compacting all your data.

                                            So, best guess: overhead penalty of duping data to maintain redundancy is recouped through compression of original data, and probably some other OBS trickery. I'm not at all sure what that trickery is exactly, though I suspect they can somehow reconstitute blocks of data??? I'm not at all satisfied with my answer but I hope to learn more.

                                            Further suppositions: I strongly suspect that there is a fair bit of overhead from all the work the OBS has to do. How much of a penalty this is I would really like to find out, but I lack the equipment (gigabit networking) to really put the hurt on them for I/O. Given my company's use case for these units, we would likely never push their I/O limitations anyway.

                                            SeanExabloxS 1 Reply Last reply Reply Quote 0
                                            • 1
                                            • 2
                                            • 3
                                            • 4
                                            • 1 / 4
                                            • First post
                                              Last post