Grey Beards on Systems - Graybeards talk object storage with Russ Kennedy, Sr. VP Prod. Strategy & Cust. Solutions Cleversafe

Episode Date: December 10, 2014

In our 15th podcast we talk object storage with Russ Kennedy Senior V.P. Product Strategy and Customer Solutions, Cleversafe. Cleversafe is a 10 year old  company sellinge scale out, object storage ...solutions with a number of interesting characteristics. Howard and I had the chance to talk with Cleversafe at SFD4 (we suggest you view the video if … Continue reading "Graybeards talk object storage with Russ Kennedy, Sr. VP Prod. Strategy & Cust. Solutions Cleversafe"

Transcript
Discussion (0)
Starting point is 00:00:00 Hey everybody, Ray Lucchese here and Howard Marks here. episode of Greybeards on Storage monthly podcast, a show where we get Greybeards storage and system bloggers to talk with storage and system vendors to discuss upcoming products, technologies, and trends affecting the data center today. Welcome to the 15th episode of Greybeards on Storage, which was recorded on November 24, 2014. We have with us here today Russ Kennedy, Senior Vice President, Product Strategy and Customer Solutions, Cleversafe. Why don't you tell us a little bit about yourself and your company, Russ? Good morning, Ray. Good morning, Howard, for giving me the opportunity to join you today.
Starting point is 00:00:52 So Cleversafe, it's a 10-year-old company. We just had our 10th anniversary this month, so we've been around for a while. Yes, we're very excited. We've been building an object storage system, which is scale out in nature, delivers tons of scalability for large customers, particularly in service provider industries, as well as large enterprises. We've been selling and marketing this product for about the last five years, been building it for 10 years now. We have a very strong customer base in a lot of different use cases in a lot of different industries. One of the things that differentiates us in the industry, obviously, object storage is new, and there are a number of participants in the industry today, but we've been
Starting point is 00:01:39 focused on this since day one. This is all we do. We're very much focused on challenging scale problems in most of our customer environments today. Getting to the 10 petabyte range, getting to the 100 petabyte range, getting to the exabyte range. We have many customers that are facing those challenges today and our solution is very well fit for those kinds of issues and those kinds of challenges that they're facing. So we're excited about the market, the opportunity. We're excited about some of the things that we're able to do for customers. And our business is just taking off like a rocket ship in the last few quarters.
Starting point is 00:02:19 I think I just heard a rocket ship takeoff here. So, Russ, tell us a little bit about the difference between object storage and normal primary block or file storage in your mind. Yeah, so I think there's really three sort of storage technologies today. Block, as you mentioned, has been around for a long time. It's kind of focused or was originally focused on a fiber channel infrastructure, you know, specific switches, specific HBAs, more of the SAN world, and more geared towards, you know, sort of the, you know, the high transaction, low latency kind of applications, databases, and other things. And there's a lot of that still running in the world today. Generally, that's the most expensive kind of storage that you have today.
Starting point is 00:03:07 And, you know, most people sort of dedicate their primary tier one applications to that type of storage. You know, file storage has been around maybe a little less time, I guess, mostly pioneered by NetApp and some of the other companies that focused on sort of NAS solutions and network attached storage, but again, primarily focused on sort of the file system structure of storage. So organizing files in sort of a structured way and a hierarchical way and allowing users to create data, to store data, to manage data, applications to do the same. But again, an underlying file system or some sort of structure associated with that. You know, Object's been around for a little less time than the other two. And really, you know, sort of the difference between Object and File is more around the way
Starting point is 00:04:02 that, you know, objects are stored and metadata is handled. So in an object storage system, objects are unique. They're stored uniquely. They're stored associated with some sort of attributes or whatever based on how the data is going to be managed and protected. And the storage system in the application had the ability to establish different types of metadata, to create different types of metadata, to manage metadata associated with the object such that those systems can grow and scale to the billions or trillions of unique objects in a given storage system. So, so they're, they're more geared towards, I'd say the scale world, you know, the, the, the, the world of, of today where, where you're trying to manage, you know, trillions of, of objects or, or even more in a single system where you're trying to manage, you know, petabytes or tens of petabytes or more
Starting point is 00:05:00 of data in a given system. So Object is really trying to solve that scale problem and the challenge of getting to large-scale systems. So you guys have environments with 10 petabytes of storage and trillions of objects? We have customers today that have over 100 petabytes of data in a single storage system, multiple customers in that vicinity, And we have customers today that are actually planning to get to an exabyte in the not too distant future. Next year, we'll have storage systems that are at the exabyte level. So yeah, it's amazing. I mean, you know, graybeards like you guys and myself have been around for a long time and, you know,
Starting point is 00:05:41 hundreds of petabytes or exabytes were were pipe dreams i remember being impressed by 50 megabytes yeah even terabytes were a good chunk of storage in one day but exabytes are real so it's it's it's amazing what's happening in the storage industry today and certainly you know the, the challenges that organizations are facing to capture and manage and store and analyze that data are tremendous. And, you know, so we're feeling very good about our position in the market and how we're able to help solve some of those problems and with proof points like customers that are actually at
Starting point is 00:06:21 hundreds of petabytes today. All right. But before we move on, there's one attribute that I generally assign to object storage that you didn't talk about, which is immutability. Okay. That, you know, it seems to me that, you know, the organizational stuff that you talk about that leads to scale is important.
Starting point is 00:06:46 But the key difference between a file system and an object store is that you can do in-place updates in a file system, and you can't in an object store. I've also started seeing vendors pitching, look, it's not a file system, it's an object store, but you can do in-place updates. And now the difference really blurs to me. Where do you guys stand on this? So I think it depends on the application's need and what sort of protocols you're using to,
Starting point is 00:07:18 you know, to store and manage the data. So if you're familiar with sort of the de facto object protocol today, it's based on Amazon's S3. Simple Storage Service, I guess, is what S3 stands for. That protocol allows you to do object updates and updates on given objects. And if you're using that protocol or supporting that protocol and your application needs it or whatnot, you are able to do that. Most traditional protocols, I would say, and our original protocol, the first one we came out with when we first launched the product as an object storage system, had that one unique object. You store it. It's stored. It's immutable. If you update it, you have to actually completely rewrite it, and you store a second version of that, and it is also immutable. The world has sort of changed a
Starting point is 00:08:11 little bit, and the need has changed a little bit, and I think some object vendors are now enabling and allowing object updates using the S3 protocol to be part of their methodology and how they're supporting data. The question I had was Cleversafe does something with geodispersion that's sort of unique in the industry as far as I can tell. Can you describe what happens there, Russ? Sure, absolutely. And I do think it is unique. I mean, others are starting to talk about some geodispersed capability, but, you know, this has been part of our product since day one, and I do think it is unique. The way our systems work, you can have storage nodes, which we call slice stores, geographically dispersed across a given region.
Starting point is 00:08:58 They can be geographically dispersed across a campus or across a city or across a country, or we even have customers that have parts of their system in one continent and the other parts of their system in another continent. Continents. Continents, yes. Continents. As long as you have a network and a network that can ship pieces of data from one location to another, you can spread your data out across these multiple regions. And we use a technique called erasure coding. We call it information dispersal. And
Starting point is 00:09:32 essentially, that technique allows you to take in data, slice it up into various pieces, add some parity associated with erasure coding, and then spread those pieces out, right, each individual slice or each individual piece to an individual storage node in a separate location and store the data there for, you know, for its permanence and persistence. What that allows you to do is set up a system that can tolerate site outage or site failure, if you will. So it's built-in disaster recovery. It allows you to set up systems to even tolerate multiple site failures. We have customers today that can tolerate at least two site failures, and still their data is available and treatable. Now, what you have to
Starting point is 00:10:17 do when you build a system like that, what we've had to do as we've matured the technology that we deliver, is make sure that you optimize the use of networking resources, of bandwidth, that you tolerate latencies across the country. I mean, between one end of the U.S. and the other end of the U.S. is about 60 milliseconds of round-trip latency. Yeah, you would think, yeah. You have to be able to tolerate that as a storage system. But the way we've designed and architected the CleverSafe system and the use of distributed erasure coding techniques allows us to do that. And you can actually run a very effective, very performant storage system using that geo-dispersed technique. And that's why we have customers today that have parts of their data in Europe and parts of their data in the U.S. It's all one, it's operating as one storage system, but allows you to, you know, to tolerate site failure. In their case, they
Starting point is 00:11:14 wanted to be able to, you know, to say that certain data only lived in certain continents and, you know, go to their users and make that claim. So. Gee, I can't imagine why somebody in the EU would want that. Yeah, there's some legal reasons associated with that, Howard. Data governance and all that stuff, yeah. But, no, that's a very important property that you pointed out, Ray, and certainly something that's been part of our approach to data management since the day one solution that we put out in the market. From your perspective, it's different than just plain replication.
Starting point is 00:11:48 It's true, almost erasure coding, you know, Reed-Solomon kinds of stuff across data and parity, right? Yeah. So it is not replication. And let me draw the distinction, because if you replicate data, that means you have to have, you know, multiple X storage systems physically distributed. You have to, you know, let's say you're going to replicate something three times. You have to have three times the amount of storage in each of your locations in order to be able to store it and manage it. So the difference in CleverSafe is our overhead, our expansion factor, if you will, is somewhere between 30% and 80%.
Starting point is 00:12:29 So for, let's say, you're storing a petabyte of data, you need storage for 1.3 petabytes or you need storage for 1.8 petabytes. Depending on your fault tolerance level and that sort of stuff. Correct. Exactly. Depending on how much fault tolerance and how much failure you want to be able to tolerate, you can design your system to be able to tolerate multiple simultaneous failures and still be able to retrieve the data. We have this tool that we use when we help customers design a system. It's a reliability calculator. And it takes some inputs,
Starting point is 00:13:07 you know, how many sites you have, how much data you're going to store, you know, what your network performance is, what your site availability is, and it will calculate the numbers of nines of reliability of the data on a permanent basis. And many of our customers today are storing data in systems that are, you know, 11 nines, 12 nines, 15 nines of reliability with less than 2x overhead. So, you know, you're getting high levels of reliability without the overhead associated with buying and storing multiple copies. You know, that's what people
Starting point is 00:13:39 want, especially at scale. I mean, it's hard to justify, you know, spending 3x the amount of storage that you need when you're storing 100 petabytes of data or an exabyte of data. Yeah, no doubt. So let's see now, at 15x, that means your data is more likely to be lost by being struck by an asteroid than by an equipment failure? Yeah, exactly. Exactly. I mean, the mean time to data loss in those calculations is in the trillions of years. So a long time after us graybeards are no longer around, the data is still going to be safe. Yeah, there's a little problem with the software, but that's a different discussion, right?
Starting point is 00:14:24 I once worked on a system that had 11 nines, but it didn't quite make it. In theory, theory and practice are the same thing. In practice, sure. So how do customers actually reference the data in this object store? Are they coding directly to S3 or are they using some other facility as a gateway? Well, so we support a number of different access methods. I mean, S3 is probably the most common and most popular that we support. We also have our own unique object interface, which we call Simple Object. It's very simple. It's designed for customers that have written their own application and they want
Starting point is 00:15:05 their applications to talk directly to the store. So it's a very simple put to store an object, get to retrieve an object, delete to get rid of an object. And we have some customers, particularly our large-scale customers, that have their own applications that use that simple object protocol. But S3 is very common. OpenStack Swift is starting to become more common. If you do need file or NAS-oriented access, we do support third-party gateways. A number of the gateway technologies that are in the marketplace today are fully supported with CleverSafe. So if you do need that kind of access, we do support those. Direct access from backup applications or archiving applications that use the S3 protocol we have integrated with some of the large backup application vendors. So many of our customers
Starting point is 00:15:52 today are using our storage systems for what I call a consolidated private cloud, if you will. So multiple applications, multiple use cases, all being stored in a single scale-out object storage system that has different interfaces, different protocols, can support different types of data. But the nice thing or the advantage that that gives you is obviously you're not replicating to protect data, and you're able to scale up and scale out to the capacities that you need, but still be able to handle the different data types. And many of our enterprise customers are definitely going down that road. So is this like a software-only solution or it's a hardware appliance or a little bit of both? It depends. We offer multiple ways to purchase and deploy the CleverSafe system. We do offer an appliance,
Starting point is 00:16:44 which is fully integrated. It's hardware and software together, and you can buy that from CleverSafe, and we fully support all the hardware and software associated with that. If you want to just procure the software, you can run our software on qualified hardware, and we've qualified a number of the major hardware vendors out there, their solutions, particularly their large-scale solutions, and we're adding more to that list every day. You can also buy and run it as an entirely virtual solution if you want to. You can run the entire CleverSafe system as a fully virtual solution under VMware, and so you can repurpose old hardware if you want to using that
Starting point is 00:17:28 model. And we've got several customers that are doing that as well. That's generally used mostly in proof of concepts or test dev kind of scenarios, but people do use it and can use it. And it gives you the flexibility and whatnot to leverage older or repurposed hardware for an object storage system. So I didn't hear anything like a service offering. Do you guys offer CleverSafe as a service offering? We do not. We do not. We have several customers, several of our customers are service providers themselves, and they purchase our technology and offer it as a service to their customers. So that's a model that we've started with and have supported for a long period of time. But we ourselves are not a service provider.
Starting point is 00:18:16 Better model for you. Yeah, we don't want to compete with our customers. Absolutely not. We definitely want to support them and make sure that they have everything that they need to deliver storage as a service using our technology. Do you have a pay-as-you-go revenue model for those people? We do. We have various pricing models, and we do enable them to grow with us. The challenge on that particular model is the hardware, and some of our hardware partners have joined us in doing that, pay-as-you-grow kind of approach.
Starting point is 00:18:49 So it's worked out for particularly some of the new sort of startups or new emerging companies in the storage-as-a-service model. It's enabled them to sort of start with a base and grow and add customers. And we like that model very well because our system scales very well. It scales very easily. And many customers are able to start small and grow up with us. Speaking of multi-tenants kind of environment, what sort of security offerings do you guys provide in this environment? So internally within the data and how it's managed is a built-in encryption model that doesn't require separate key management. So every piece of data that's ingested into a CleverSafe system goes through
Starting point is 00:19:43 an encryption process. It's done using a randomly generated encryption key, and it's done on a segment basis. And as I mentioned before, data is then erasure coded and sliced into pieces, and those pieces are distributed around a network of storage nodes. What that means is if you don't have the predefined subset of pieces or slices to retrieve the data, you don't have the predefined subset of pieces or slices to retrieve the data, you don't have enough information to actually get the encryption key to decrypt the data to return it back in its original form. So let's say you had data spread across three different locations in three different cities, say, if you will, and someone
Starting point is 00:20:22 were to compromise one of those locations, generally it's engineered such that there's not enough data in that individual location to actually retrieve the encryption key to decrypt the data to return it in its original form. So the data itself is 100% secure. We also use TLS in transferring data between nodes in our system. So data is encrypted as it's in flight. It's also encrypted as it's stored using that encryption method that I described. So 100% the data as it transfers through our system and as it's stored in our system is encrypted. So it's protected from that aspect. We also enable our customers to decide what sort of authentication methods they want to use in order to enable and ensure that their users and
Starting point is 00:21:13 their customers have access to the data. So we integrate with a variety of different authentication methods, Active Directory, LDAP in general. We use PKI if they want to, you know, to use the private key initiative authentication methods. So we give customers a variety of ways to secure the data and make sure that the data is protected. Okay. Okay. So we got security. So tell us, so in your environment, you mentioned accessor stores and then slice stores. And what are the purposes of those? So slice stores are storage nodes. They do two things. One, they store slices of data.
Starting point is 00:21:51 So out of the distributed erasure coding model, slices of data are actually stored on those storage nodes. It's just a bunch of drives with a server and some memory and network interface. That's what a slice store looks like. The other function that they perform is the data integrity function or the rebuild function, if you will. So one of the processes that runs in CleverSafe in the background is this rebuilder process, and there's two functions to it. There's a scanning process and an actual rebuilding process.
Starting point is 00:22:25 The scanning process checks the integrity of every single slice that's stored in the system constantly. And if it finds a slice that's missing for some reason, can't be read because there's a drive failure or some spot on the disk that can't be read, if a bit is flipped because a bit rot or something like that, it detects that that slice is unreadable and it will rebuild the slice from the contents of the other slices that exist in the system.
Starting point is 00:22:50 As long as you have a threshold number of slices for a given segment of data, you're able to rebuild any of the missing slices. And so Rebuilder is a function or a process that runs on the slice store nodes in the background. Access to the system is provided through what we call an accessor device or an accessor node. And there's various models for the accessor nodes. First of all, they're stateless devices. Data is not stored on them. It's just passed through them.
Starting point is 00:23:16 And they're responsible for the encryption process that I mentioned, the distributed erasure coding process that I mentioned. They're the devices that actually tune the performance of the network or validate the tuning of the performance of the network so that slices can move efficiently across the various nodes and the various locations and whatnot in the system. So the access device is responsible for giving access to applications for storing data onto the CleverSafe storage system. And they publish or support the various protocols that we support.
Starting point is 00:23:53 So our own simple object protocol, the S3 compatible protocol, the Swift compatible protocol, the HDFS, Hadoop distributed file system File System Compatible Protocol, are all supported natively by accessor devices. And again, you can deploy them in a variety of different fashions. You can deploy them as virtual machines. You can deploy them as physical appliances. You can actually deploy them as an application running on your application server, so you don't have to talk to another device in the network. You can talk directly to the storage through your application. So you have a variety of different deployment options for accessors. And again, they're stateless. So the way most customers deploy accessors is they either choose a deployment model, either physical appliance or virtual appliance and whatnot.
Starting point is 00:24:40 And then they bring up the accessor nodes, and then they load balance the network traffic across those accessor nodes so that they can leverage the performance of them individually. They can leverage the scale of them by adding more if they need more performance. They just use some sort of load balancing technique to distribute the workload across that bank of accessors. And again, if one fails, the other ones can take over. There's no issue or no problem associated with accessors being single points of failure. So you mentioned HDFS. I mean, how does CleverSafe work on HDFS cluster? I don't understand that. So two ways. One, we are a target. If you want to store HDFS data, you can do that by just pointing HDFS or Hadoop nodes to CleverSafe as a target and just store that data just like you would on a standard Hadoop cluster. So instead of having local storage in the nodes that run MapReduce? Absolutely. Correct.
Starting point is 00:25:46 And generally what customers use us for versus a standard node that's running MapReduce is scale. So if you want to get to two petabytes or three petabytes or 10 petabytes of Hadoop data, it's a little challenging to do that by using just standard Hadoop nodes and Hadoop file system in general. So you can use HDFS to store. But I'm Rackface space, and I can fill thousands of racks with servers. Sure you can. But then Hadoop natively out of the box requires 3X replication, so you're going to have to buy three times the amount of storage right whereas you can get higher levels of reliability with clever safe and our distributed erasure coding at 1.3 to 1.8 times within the same data center you know even within the same data center yeah exactly even within the same data center so your cost and your economics are are
Starting point is 00:26:39 are there you get the distributed nature of our our solution if you want, which you don't necessarily have with a standard Hadoop cluster. And oh, by the way, if you want to actually run MapReduce on our storage nodes, you can do that as well. We support the ability to run natively MapReduce functions on our storage nodes. So you can actually store the data and then analyze the data directly on the storage system itself. Wait, wait, wait. You're running Hadoop MapReduce on slice stores? Yes, exactly.
Starting point is 00:27:11 Exactly. Wait a minute. Does this make any sense? I mean, the reason you'd run it, well, okay. So the reason you would want to run these things is because the data is close by, but the slice store has this intrinsic geographic or distribution capability. I don't know. I don't know. I don't know.
Starting point is 00:27:28 Yeah, but when you're not ingesting data, the slice stores aren't all that, have a lot of CPU available for other things. It's got a lot of process. Howard's exactly right. It's got a lot of excess process capacity. And again, think about it at scale, right? Most Hadoop clusters I think today are, you know,
Starting point is 00:27:41 they max out at about a petabyte. If you've got, you know, let's say 10 petabytes of data or 50 petabytes of data that you want to be able to analyze maybe over time, you want to archive it, but then maybe go back and analyze it sometime in the future because something's changed or some event has happened or something like that. It's a perfect solution for that. It eliminates the whole, take the archive data, roll it into Hadoop, run MapReduce, throw it away, do it again next month. Exactly. Exactly.
Starting point is 00:28:07 It's all resident. It's all permanent. You can go analyze it again. It would be interesting to see how much of the average Hadoop cluster's time is spent with data ingest and data destruction. Well, we see it as a productivity improvement, especially for customers that have a lot of data that they want to be able to analyze at some point in time. Right. Because you're not moving the data. You're just running your MapReduce jobs whenever you feel like doing it. Speaking of performance, is there like an SSD version of this? I mean, can you plug in SSDs
Starting point is 00:28:41 in all these locations? We do have a solid state appliance that we market, so you can run an SSD version of a Slice store if you want to. It's generally lower capacity, obviously, and higher performance. So you'd want to set aside a certain workload or a certain type of data that you'd target to SSD. But I'll tell you, that's an area where we're definitely focused from a development and future delivery standpoint, is integrating multiple tiers of storage, adding policy and capability to move data around across a variety of tiers. That's one of the things that many of our customers are asking us to do. And so we're definitely focused on that as
Starting point is 00:29:30 enhancements. But you can today run a solid state version of this if you want to. It would cost you a lot, but you can. It seems strange to me because, you know, the real strength is the dispersal coding. So if I'm running 16 of 20 erasure coding, every time I want to read a block, I've got to ask the guy across the WAN for some of his slices. Yeah, I would say most— That introduces so much latency. So what I would say, Howard, to that is you wouldn't geodisperse a solution. You would put a solid state version in one data center, if you will, and then have a copy of the data that would live in that solid state erasure-coded system. So you're not traversing the WAN to get data. There is a little bit of overhead, obviously,
Starting point is 00:30:22 with the erasure coding that you do have to factor in. Yeah, but in cluster latency, we're talking microseconds, not milliseconds. Yes, correct. Correct. So that is true. And when we have deployed solid-state versions, it is, like I said, in a single data center, whereas the high-capacity solutions are the ones that are geodispersed. So what does an exabyte of data look like, Russ?
Starting point is 00:30:50 I mean, Jesus, we're talking football fields of the storage here. Well, let's see. What's drive capacities today? Six terabytes going to eight terabytes soon. You can do the math. How many drives is that? And then with a little bit of overhead, let's say, you know, 30%, let's say, you know, maybe average 50% overhead. You're talking, you know, you're talking thousands and thousands of drives and, and hundreds and maybe
Starting point is 00:31:15 thousands of nodes in multiple locations with racks and networks and, and, and, you know, switches and all kinds of stuff associated with it. Yeah, it's a big system, no doubt. Lots of physical architecture associated with an exabyte of data, for sure. So are... Thumbs now? Because if I've got that all-flash one in one data center, can I then replicate it across the pond? You absolutely can.
Starting point is 00:31:43 You absolutely can. That's a very popular use case for us. So storing a local copy and a mirrored copy that's geodispersed, that's a very popular use case for us. Yeah, it gets you better reliability than three copies and now you're at 1.5 or 1.7, not three times? Correct. Correct. So in that situation, you would be both replicating as well as geographic erasure coding? Is that how I read that? No, in that situation, so you'd have two independently, what we call storage pools, independent pools of storage, maybe one flash-based and one, you know, high-capacity
Starting point is 00:32:26 spinning media-based. We have this feature called vault mirroring, which vaults are logical targets for us, like volumes in a, you know, in a sand system. Vault mirroring, what that says is for every object that's written to a particular vault, write a mirrored copy of that object. So you write one copy in your local storage pool, which is, let's say, flash-based. It is erasure coded, but it's all in one physical location. And the second copy goes to a geo-dispersed vault, which is across three different sites. In that case, you have a little bit more than 2x overhead because you've got two copies, right? So you don't have, you know, 3x or 4x or 5x overhead. And your reads, you know, would likely come from the local copy. So your performance on read would be very, very fast.
Starting point is 00:33:19 Yeah, yeah. But if I had a read failure, it would automatically go to the other one? Absolutely. Absolutely. Yep. Automatically. So you mentioned a stateless accessor store. What does that mean? Configuration data, all that stuff must come off the network then? It does. It comes out of the system.
Starting point is 00:33:36 So all the configuration, all the what we call the registry, which is sort of the roadmap or the map of the infrastructure, if you will, is all distributed across the network. So you can bring accessors on and off as you need them. Let's say you have a burst in workload over a given period of time. You want to add more accessor capacity. You can spin up physical accessors or virtual accessors. They're essentially stateless devices. No data is actually stored on them.
Starting point is 00:34:03 They get their information from the network, from the system, and they come up. They know which vaults they're supposed to talk to, and away you go. When you're done, you spin them down, and the system continues to operate. That's one of the design tenets that we have and have had since day one is that systems go in. They're intended to operate 100% of the time forever. And, you know, we've had several customers that have been customers for many, many years and have never taken their system down for an upgrade, for adding capacity, for actually physically moving the system from one location to another. You can do all of that with the system
Starting point is 00:34:45 live and operational using erasure coding. As long as you maintain access to a threshold number of slices of the data, it can continue to operate even if certain elements are down. So if you're moving from one location to another, you'll do it over a period of time, but you'll take a couple of storage nodes out, physically move them to the next location, bring them back online. This rebuilder process detects, hey, I wrote some data that I need to make sure is now on the new nodes or the nodes that have been physically moved, and it will catch up in the background for you. So all of that is possible using our system. Can you run this sort of thing in the cloud? I mean, could you effectively fire up an accessor store, an S3 or something?
Starting point is 00:35:29 An S3, no. Amazon.web services and things? If you wanted to, sure. We've had some customers ask about doing that capability. No one's done it to date, but it is possible. Certainly other public cloud providers, there are many of them out there, and you could certainly do that if you want to. Again, the entire thing could be run as a virtual instances, so you could spin up everything in a public cloud provider if you
Starting point is 00:35:57 wanted to. That's interesting. An exabyte of EBS would add up, though. Yeah, it might be a bit of a challenge. Is that what you're saying? An exabyte of EBS would add up, though. Yeah. It might be a bit of a challenge. Is that what you're saying? Yes. It certainly is a challenge to your pocketbook, for sure. Well, I was thinking you could put the Excessor stores in AWS, and you could reference them from anywhere, and you could have the Slice stores, as it were, be your own physical on-prem storage in numbers of locations and stuff like that.
Starting point is 00:36:25 I don't know. Certainly possible. Not sure why you'd want to do that unless you wanted to run the applications themselves in AWS. It's also possible. Yeah, I think you've got the latency one way or the other. Well, again, if you're designing a system for very low latency, very high transaction, this is probably not the solution for you. There's better ones out there. But we see the world sort of bifurcating.
Starting point is 00:36:53 There's performance-optimized storage and there's capacity-optimized storage. And, yes, you have to have performance in the capacity world. But we see what we're doing is being able to solve sort of the bulk of data storage needs, if you will, out there. But we're not going to try and become the world's fastest Oracle database storage system. That's not true. And the truth is, for capacity-oriented storage, performance mostly means throughput, not latency anyway.
Starting point is 00:37:22 For the most part, yes. For the most part, you're correct. Yeah, it's funny. Some of our fellow analysts have proposed that RAID is dead and we're going to have to move to the kind of erasure coding you guys do for primary storage. And I really have a lot of trouble seeing that.
Starting point is 00:37:42 Just the latency that it would introduce for things like a small right in the middle of a stripe just i think yeah we're certainly not saying that i mean there's plenty of obviously data that's being generated if you're talking about exabyte storage systems that they'll keep us busy for a long period of time. So we're not saying we need to go in that direction at all, Howard. And I agree with you. I think there's very good storage systems for that kind of use case and that kind of need. And absolutely, you know, the need is out there in the world, but there's, you know, what we do focused on, you know, sort of the capacity needs and whatnot. And I think we've got a very good solution for those kinds of challenges.
Starting point is 00:38:28 Yeah, historically the challenge with something like object storage has been having a protocol that everybody, you know, an industry standard protocol and stuff like that. And how does this play into that to some extent? You know, so you've got NFS and CIFS SMB and block storage. All those protocols have been defined by committees, and they're out there, and they're standardized, and they're certified and all that stuff. Object storage is a different animal altogether in that respect. Yes, you're absolutely right.
Starting point is 00:38:57 I'm hoping, I'm hopeful that it will mature into something along those lines. I'd say the de facto standard today is Amazon S3. Most of the applications in the world that are speaking native object are conversant with Amazon S3. So it's sort of the de facto standard, if you will. But I think the industry in general, the vendors that are object storage vendors
Starting point is 00:39:21 and the industry in general does need to do some work around standardization and getting to something that everybody agrees is the right protocol. And the same thing happened when, you know, when NFS became, you know, a standard. Same thing happened when SMB became a standard. There was emergence and then there was, you know, everybody got on board and blessed it and said, this is what we're going to support going forward. The same thing needs to happen in object storage. Hopefully it'll be soon. I agree with you.
Starting point is 00:39:51 And when that happens, I think more and more applications will then be object conversant. And that will be a good thing for customers, I think. Yeah, I think we put too much weight in du jour standards, though. De facto standards work fine i mean the truth is no you know we were using smb for many years before anybody actually wrote down a spec for it right um so you know and cdmi is kind of the the example of what happens when you start writing a spec before you let the market figure out what the use cases are.
Starting point is 00:40:29 Yes, yes. So now you have a standard that nobody supports. I would agree. I would agree. But the challenge, if one particular organization, particularly a vendor uh sort of owns the de facto standard is they can change it right they can they can do whatever they want with it in that manner and then you know amazon can do whatever they want with s3 right just like microsoft you know came out with smb 2.2 and smb3 um and you know the truth is most people just continue to support the old version. Yeah, exactly.
Starting point is 00:41:07 Although there are people moving to SMB3 and that sort of thing. I mean, obviously there are advantages there, and Microsoft has made it a public standard, so that's also helped. Right. Yeah, I think that the world will evolve that way as well on Object. And I think once it does, I think Object will become more and more mainstream. It's it does i think uh object will become more and more mainstream it's already becoming mainstream but it'll become more and more mainstream yeah i just think we should use you know take smb smb more as of our of our example is then nfs where you know nfs4 and nfs 4.1 those specifications have gotten so
Starting point is 00:41:42 complicated and overwhelmed with special use case add-ons that nobody's actually implementing them because they're too hard. I know one customer that's at Microsoft, of all things. Yes, interesting. Yeah, yeah, yeah. I mean, what do you guys think of the old OpenStack initiative? I mean, that could be a way to go as well, right? An open source standard. Yeah, it's getting a lot of press.
Starting point is 00:42:14 It's getting a lot of interest. I mean, most of the storage vendors are looking at it seriously. I mean, a lot of the cloud vendors are looking at it to a large extent. So, yeah, I think it's gotten a lot of good good momentum going with it it makes it makes a huge amount of sense you know from the storage business point of view you know if you're a block storage vendor writing a cinder driver is easy and you know if you don't do it then it makes you look like you're behind and you know and VDI, it's an interesting use case that incumbents don't have a good answer to. So everybody else flocks to it. Whether it turns out to be huge or not is a whole other story. Yeah. But I do agree. It's certainly the popularity is strong out there. And
Starting point is 00:43:03 who knows? It may become more and more the standard. Well, I mean, certainly the latest couple of revs of OpenStack are coming a lot closer to being ready for enterprise prime time than the early ones. Yep, I would agree. I would agree. So where does something like Lustre and GPFS and those sorts of things fit into this framework from your perspective? I mean, they've historically been, you know, scale-out file services kinds of things. Yeah, and historically been, you know, sort of focused on, you know, the HPC, you know, model, if you will, or, you know, those kinds of things. You know, they certainly have their place in the world. You know, I think Object is more general purpose, certainly more, you know, capacity oriented than those guys.
Starting point is 00:43:53 And, you know, we certainly see, you know, customers that do have a need for, you know, for those scale outout file systems, but there's plenty of need out there for, you know, for large and structured general purpose data like, you know, like CleverSafe supports. And, you know, I don't see it, you know, sort of entrenching on object in general. I do see it has a play and, you know, it seems to be, you know, like I said, more focused on the HPC environment. Well, it's much more latency sensitive. Much more for latency sensitive applications. You can talk about a clustered file system
Starting point is 00:44:32 as scale out, but it only scales out as far as the back-end storage system scales out. Sure. When you hang 27 front ends off a VMAX, the VMAX still is the limiting factor. Correct.
Starting point is 00:44:48 Are you talking VMAX 3 here, Howard? I'm sorry. Go ahead. No, I agree with you. I think it is certainly more oriented towards low latency kind of applications. And there's a need for that. Absolutely, there is a need for that. Absolutely, there is a need for that. But there's also a need for general purpose, you know, scale out object storage systems that are not, you know, necessarily focused on low latency applications. You know, I would have thought object systems would have been a natural for HPC model.
Starting point is 00:45:14 But the issue there is not the, I'm sorry, the performance as much as it's the scale requirements that are critical. Well, we've come to just about the end of the conversation here. Howard, do you have any last questions for Russ? No, I just wanted to comment that I found it interesting that we've reached the point where we're talking about object storage as more general purpose than GPFS, where GP does stand
Starting point is 00:45:38 for general purpose. Yes. Yeah. I'm actually pleased we've reached that point where two or three years ago when we talked about object storage it was definitely this is the niche solution
Starting point is 00:45:54 it's a hell of a niche if it's a niche solution well I mean they were big niches it's not the niche you put your Capitamonte miniatures in it's the one you put your Capitamonte miniatures in. It's the one you put the full-size copy of Michelangelo's David in. But it's still a niche. Or at least was, yeah.
Starting point is 00:46:16 It's certainly becoming much more mainstream than niche. The application front side has gotten mature enough that now we can, you know. Absolutely. And if those folks that think that, you know, sync and share is going to replace file services for end users, then it really does become general purpose. I agree with that. Russ, any last comments? No, I've enjoyed spending some time with you guys. Great questions, and I certainly enjoyed the dialogue and discussion,
Starting point is 00:46:49 and look forward to catching up with you in the not-too-distant future again. Well, this has been great. Thank you, Russ, for being on our call. Next month, we'll talk to another startup storage technology person. Any questions you have, let us know. That's it for now. Bye, Howard. Bye, Gray. Bye, Russ. And. Bye, Howard. Bye, Gray.
Starting point is 00:47:05 Bye, Russ. And thanks again, Russ. Okay, Ray. Thanks. Until next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.