Grey Beards on Systems - GreyBeards talk all-flash storage with Vaughn Stewart, Chief Technical Evangelist Pure Storage

Episode Date: September 16, 2014

Welcome to our 12th monthly episode where we return to discussing all-flash storage arrays, this time with Vaughn Stewart, Chief Technical Evangelist for Pure Storage. This our third all-flash stor...age array vendor that the GreyBeards have talked with since the start of the podcast, not counting server side flash and scale-out/hybrid storage systems. Vaughn has had a long … Continue reading "GreyBeards talk all-flash storage with Vaughn Stewart, Chief Technical Evangelist Pure Storage"

Transcript
Discussion (0)
Starting point is 00:00:00 Hey everybody, Ray Lucchese here and Howard Marks here. Welcome to the next episode of Greybeards on Storage monthly podcast, a show where we get Greybeards storage and system bloggers to talk with storage and system vendors to discuss upcoming products, technologies, and trends affecting the data center today. Welcome to the 12th episode of Greybeards on Storage, which was recorded on September 5, 2014. We have with us here today Vaughn Stewart, Chief Technical Evangelist of Pure Storage. Why don't you tell us a little bit about yourself and your company, Vaughn? Hey, guys. Thanks for having me on the show.
Starting point is 00:00:45 First off, what do you mean September 5th? It's September 6th where I'm at. All right. All right. You're in the future. You're in Australia. We're in the U.S. I guess we'll have to deal with that somehow.
Starting point is 00:00:56 Hey, look, it's great to be on the show. You know, I know we all just caught up with one another just a few weeks back at VMworld, and I appreciate the opportunity to come in and share with you a little bit about what we're doing at Pure Storage. I guess, what would you like to know? Well, you can start with what is it? I mean, obviously it's a storage device of some type. Sure. So Pure Storage was founded on the vision of enabling flash storage to be broadly adopted throughout data centers. Currently, we ship a product. It's the FlashArray 400 series. This is comprised of entry-level, mid-tier, and a high-level storage array, and they scale it both in performance and capacity as you go through those tiers.
Starting point is 00:01:56 The systems offer block-based storage access, so iSCSI or fiber channel, with the bulk of our sales being fiber channel. No kidding. That's good. So SCSI or Fiber Channel, with the bulk of our sales being Fiber Channel. No kidding. That's good. Yeah. And the design of the systems is 100% Flash. And through a product, what we call, or a feature set we call Flash Reduce, we provide five forms of data reduction technologies that allow us to invert the price premium of flash storage and offer it at a price point on par with that of spinning disk, high-performance spinning disk, I should say. And thus, our target market is really tier one workloads, be it OLTP databases, virtual infrastructures,
Starting point is 00:02:43 and virtual desktops being the three primary areas of focus. So it's an all-flash array? Correct. And is it – so you mentioned five forms of data reduction. Could you expand upon that or expound upon that? It sounds kind of like build strong bodies 12 ways. Sounds a little marketing-y to me. All right.
Starting point is 00:03:07 Well, you know, we still want to know what the hell it is. Well, look, you know, storage systems have had forms of data reduction built into them for quite some time. And they've slowly trickled from the backup space, you know, a la data domain. But even prior to that, right, we were compressing data as we were putting it on a tape. So we spun from tape backups to disk backups. I was the one who brought NetApp's block-level deduplication out from the backup use into the production use with server virtualization back in, what was that, 07, 06, 07? Yeah, yeah, yeah.
Starting point is 00:03:44 You know, so we learned a lot from that, and that's actually part of what drew some of my interest a few years back with Pure Storage. Different types of data, or different types of data sets to be more specifically, have different responses in terms of the benefits or gains based on the type of data reduction technology that you apply towards them so pure storage groups our collection of data reduction technologies and we refer to them as flash reduce and so basically we have an inline form of pattern removal pattern removal is the process that some applications or systems or i should say patterns is how some applications or
Starting point is 00:04:23 systems try to subvert thin provisioning mechanisms. So we have pattern removal, so to keep all data on the system thin. Second, we have an inline form of data deduplication. It's very granular. It has a 512-byte granularity, which allows us to provide the greatest benefits or returns on capacity across a data set. And the 512-byte granularity we can get into a little bit more if you'd like, but granularity block size really demonstrates its strength over the long haul as data blocks start to get dirty and guest file systems of the objects that are being stored by the array.
Starting point is 00:05:08 We also have a form of inline compression. And so compression and deduplication kind of at a high level service to two different markets. Deduplication sees strong data reduction returns on unstructured data sets or where you've got file commonalities such as the binaries within virtual machines, whereas the strong suit for compression is really within the application space. So it's three forms. The fourth form is what we call deep reduction, which that is a marketing term.
Starting point is 00:05:44 And deep reduction really is a second form of compression, one that's much more aggressive than what we do inline. And we pair that with the back-end system process and garbage collection so that as we are having to manage the data over its long period of time, we're actually returning additional capacity savings to the array. And that deproduction provides another 2x capacity reduction over the original compression on the inline element. So is that application-aware filters like Ocarina used to do, or is that bigger dictionaries in the compression algorithm,
Starting point is 00:06:23 or is that just magic? Well, yeah. Step A, step B, and in between, this is where the magic happens. So the inline compression is, it's best to look at it this way. So compression has a correlation between the amount of CPU that you can throw at the algorithm and the data reduction rates. And so what you see across the industry on inline compression is a very moderate form, usually one of the LZs. It could be LZ4 or, you know, we're LZO. And our LZO is, if you think about it in terms of aggressiveness, let's say it's set at about one-third aggressiveness
Starting point is 00:07:05 in terms of its max. And this gives you the ideal mix between data savings on ingest as well as still being able to maintain some millisecond response times externally. When we transition to deep reduction, what we're actually doing there is at the time of garbage collection
Starting point is 00:07:25 where we're having to move valid data cells because we want to erase a block that has some invalid data cells. When we go to move that data, we actually apply a more aggressive ratio of the LZ0. So say going from one-third aggressiveness to about two-thirds of max settings. And then we also apply a very aggressive proprietary form of Huffman encoding. And the benefit of this approach is obviously to net customers more capacity on the system during a normal operation that every flash device has to do, right? Every SSD has to garbage collect. But by being able to layer these two and do them post-process,
Starting point is 00:08:10 which is funny because some vendors knock this, but to actually do it post-process is to be able to provide a form of data reduction that's just really too taxing to do in line. And so I don't understand some of the FUD that throws around. But, you know, we've all been around the storage block. It's kind of just the state of the union. Sometimes you just find something that's different.
Starting point is 00:08:31 Yeah. Different or, you know, storage vendors are like that you wouldn't do in Candide's best of all possible worlds, where 156 core processors cost $10. It's funny because it's, you know, alien technology across vendors is almost like,
Starting point is 00:08:50 you know, it brings out the xenophobe and everyone must crush it. But the reality is that system architectures are designed with trade-offs, right? And so you design your system for a type of workload or a type of functionality. And, you know, while many of us think Intel processors are these magical beings that can do everything, and they can, they just can't do everything simultaneously. Right. Yeah. So you got up to four. What's the fifth and final deep reduction technique?
Starting point is 00:09:18 Yes, as we roll the tangent back in. We don't always do that. Yeah, you're right. We consider that a success. Yeah, okay, thanks. And we digress again. So the fifth is to actually not write the data at all. So as you know, these data reduction technologies come together to reduce the cost of the system, but
Starting point is 00:10:12 probably what many people don't realize is the data reduction technologies actually increase the resiliency of the flash media because the flash media itself has a limited life cycle. Every byte you don't write is one write erase cycle that FlashCell doesn't have to go through. Exactly. So you mentioned, so the fifth is actually like a copy on reference snapshot kind of capability? Is that what you're saying? Yeah, think of it as, I'll put this in generic storage terms,
Starting point is 00:10:44 think of it as a pointer-based copy. I've got a set of data blocks on my persistent layer, and someone wants to make a copy of it. Instead of writing the data a second time and or then having to deduplicate the data a second time, why don't I just make another metadata reference pointer to the data and now have two instances of that object versus one? Yeah, so if a write comes down for one of those instances, then you create an old version of the block or something like that? Is that how it plays out? So the two predominant use cases we see are in virtual machine cloning via hypervisor or database clones. And so when I'm receiving a request, let's say from VMware, leveraging their XCopy API set, it's saying take this block range that I understand to be a template of a virtual machine and make a second copy of it. And we respond back with saying, okay, well, here's a new set of external logical block
Starting point is 00:11:49 addresses that make up your second reference, but they're just metadata reference pointers internally. You know, the provisioning, the metadata reference, cloning an entity on a storage device by additional metadata reference pointers is really the inverse of taking two identical sets of data and collapsing them onto each other through deduplication. It's two sides of the same coin. I've reached the point where I think that, you know, what I've actually started calling, you know, what you guys are doing, VAEI unload.
Starting point is 00:12:24 You know, we're not offloading it. We're offloading the work. We're having the array avoid the work altogether and put in metadata pointers. As far as I'm concerned, that's table stakes for a modern storage system. Yeah, anymore. I think I would actually take that a step further and say, and I actually just wrote a blog post on this that I think is going over to InfoWorld. I don't know if I can say it.
Starting point is 00:12:47 If I can't say that, bleep that out. No, go ahead. It's okay. I think storing your data in space-efficient constructs is the new norm. Yeah. And there's a lot of nuances between how those technologies work and operate. But if you look at the data growth rates that are occurring, I don't know how we're going to reverse the trends any other way outside of storing data, leveraging the CPU cycles to store the data in a more dense format.
Starting point is 00:13:16 Yeah. Well, and the greater overall efficiency of the system, especially when we start dealing with constructs like VVOLs. If you don't have metadata that can handle dealing with things at that granular level, boy, is that implementation going to stink. Oh, and so don't even get me started on the VVOLs. I mean, we've been waiting roughly three or four years. And I think as an industry, we are really hoping for this payload to deliver
Starting point is 00:13:48 on its promises. But if you look at something as simple as the copy offload that's available today with the clustered file system in VMFS, there's still an intermediary there in VMFS. It's gating the transfer rate of the IO. And when we actually move to vVols, there is no more gating. It's gating the transfer rate of the I.O., and when we actually move to vVols, there is no more gating. It's a direct handoff between a hypervisor request and a storage controller's response. So you'll actually get even greater performance with vVols. Take that then a step further, and we'll actually get to see T10 unmap at the guest layer, right, in terms of the actual virtual machines operating system, you know, wow, you know, to finally be able to free ourselves of the garbage that's accumulating inside of file systems of virtual machines on data that's no longer required. Now, there's still some tap dancing that's required throughout the ecosystem to make
Starting point is 00:14:37 sure that data is not sitting in operating system constructs like, you know, temp spaces, either system or user or recycle bins, but you can handle that through different policies, etc. That's interesting. That's interesting. I was going to say, one of the ways we're going to deal with this data growth capacity issue
Starting point is 00:14:58 is that Seagate just came out with 8TB disk drives. Were you guys at Flash Memory Summit? One of the vendors announced 14TB SSDs that fit in a 2.5-inch at Flash Memory Summit? One of the vendors announced 14 terabyte SSDs that fit in a 2.5 inch form factor. What? For 2016. 18 to 24
Starting point is 00:15:13 months out. I mean, today you're seeing a handful of vendors really push the tier one storage market to say Flash is available at a price point of Tier 1 disk. And this is HP, EMC, Pure Storage is an example. These are the folks that are leading this change.
Starting point is 00:15:34 When you get a 14-terabyte drive and a 2.5-inch drive form factor, that means the Tier 2 market's next. It's two years out. Oh, come on. Now you're not going to go after Tier 2. I think Tier 2 is going to happen. There's different scales in terms of system, access protocols, etc. But I think Tier 2 comes quicker than most people
Starting point is 00:15:53 are expecting. Huh. You know, I was at a dinner at... If the disk drive vendors miss their targets on bit pattern media, if Moore's Law holds and crider's law doesn't and and there is the opportunity because there's that big technology shift that we haven't seen them try yet um i could see that happening you know clearly you know in my old home of new york
Starting point is 00:16:20 that'll happen pretty quickly because the floor space costs are going to be an issue. Because we've already reached the point where a two terabyte, two and a half inch flash drive is denser per cubic inch than that eight terabyte Seagate three and a half inch drive. Yeah. Yeah. It's funny. So we have conversations with some very large customers. And when you get into areas where there's just dense population, so it could be the financial area of New York, it could be London, you can be sitting here where I'm at in Sydney, Rackspace or data center resources more specifically, the power, the cooling, the rack space, they're at a premium. And in many of these locations, it costs more to just power a storage array annually than it does to acquire a storage array. And Flash is the way that you invert that trend. It allows you to take whatever the data capacity is of a data center today and extend it tenfold or twentyfold into the future. I don't believe it. Look, I didn't say tape was dead.
Starting point is 00:17:31 Thank God. Jesus. What's with tape? I never talk about tape. Everybody else talks about tape. I never talk about tape. God damn. I'm sorry. Getting back to the world here. So, all right, so Pure Storage is data reduction with all flash and all that stuff. I don't think I've seen any SPC1 benchmarks with Pure. Is there something here I'm missing? Well, there's much more to the system architecture outside of the efficiencies,
Starting point is 00:18:02 but that is the disruptive economics element. Without it, we don't have a conversation. We're sitting in the camp of tier zero. We are kind of at an interesting point with the SPC benchmarks because, to the best of my knowledge,
Starting point is 00:18:17 they don't accept any results where data reduction technologies are a part of the system. The question then becomes, should we not participate in SPC benchmark and invest our resources in other areas, or should we complete an SPC benchmark and know that it can't be certified? Well, yeah, my opinion, I mean, you could go in with an SBC benchmark and, in my mind, I guess, can you turn off your reduction capabilities? You can't adjust any of it.
Starting point is 00:18:56 That's the beauty of an all-flash system. There's no knob tuning. You can't adjust it. You can't turn it off. No. Yeah. So if you could turn it off, you could go in there. It would be a much higher price and you could effectively put words around the fact that with a
Starting point is 00:19:11 5x data reduction capability, it's going to be much cheaper. But if you can't turn them off... Then you'd be generating test data that doesn't reflect the real world. This is how fast it is with this feature turned off, but nobody ever turns it off. Yeah, and you bring up a really good point when it comes to benchmarking or testing these systems. I mean, IDC has published a report that gets referenced quite extensively within the market around testing an all-flash array. It's got some great information and guidance within the report,
Starting point is 00:19:50 but it also has some what I'll call antiquated views on how to test the systems. And we and other vendors have been reaching out to IDC, and to their credit, they've taken this feedback and are working on an update to their all-flash-ray testing methodology document. The way I look at the world is a third of the customers you meet will swear by testing with synthetic workloads. Another third will swear by only testing with real data. And there's this remaining third that's kind of undecided. But what we see spawn from this IDC test report is this big surge on synthetic workload testing and particularly characteristics that aren't seen in the real world. And again, it's very interesting because, as I said, there's this engineering centricity
Starting point is 00:20:41 within our profession that some people think that data, while having an extremely low probability, is highly relevant to their decision process. And others view it the opposite way. What about Exchange Solution Reviewed Program, ESRP? You could submit a solution there. There's not a lot of flash solutions in ESRP results. I was attributed that to maybe some characteristics of OLTP workloads and disk solutions. Yeah, we've seen some pretty significant data reduction results with some customers who've put Exchange onto Pure. I don't know where we stand within the Alliance team. I know we're doing a ton of investments within Microsoft
Starting point is 00:21:32 spanning from integration points, such as we just released a VSS provider within Purity Operating Environment 4.0, but also going up the stack into points of integration, certification within the platforms. So I don't know where we're at with the Exchange benchmarking, but I can follow up with you on that offline. Your deduplication model should work quite well for Exchange.
Starting point is 00:21:55 Exchange is a difficult data set to reduce because it's constantly defragging the database. And if you use a fixed chunk size, things get very strange. Yeah, so you're right. So as well as the attachments, or I should say the Microsoft Office files, in the modern versions, all those files are actually compressed formats. Right. So variable block size, which I know I said 512 byte granularity, but the variable block size that is enabled by having 512 byte granularity,
Starting point is 00:22:33 as well as multiple forms of compression, allows to reduce those files, office files, rather well. But then if you also look at how Exchange stores data, which is there is no form of single instance anymore. And in fact, significant redundancy around availability groups. We can collapse those environments quite aggressively.
Starting point is 00:22:55 What I think more of an interesting construct when you actually go beyond the technology is when you say I have an all-flash array, you tend to elicit a response from 80% of the market, which is, ooh, flash. That's only for some mission-critical workload.
Starting point is 00:23:11 I don't know if I need that within my data set. And yet there's this smaller subset of the market, and it's a growing subset that is starting to recognize, if I can take the economics out, which, again, I think is the play that you see occurring with HP, EMC, Pure, etc. If I can take the economics out and I've got to spend my budget on storage regardless if it's disk or flash, flash with sub-millisecond latency actually makes the use or the quality of the experience with every application better. And so what we see with customers who have moved Exchange onto Pure is this
Starting point is 00:23:48 Google-like interface experience when you're in your Outlook client. I start to search and I'm getting immediate responses back. My troving of my archives is no longer a waiting task that you avoid. It's an immediate response, much like what we receive. How many cores do you have on your desktop? Yeah. Not enough. Because doing searches in Outlook against the SSD in my desktop isn't a Google-like experience.
Starting point is 00:24:18 That's because of that cheap SSD you've got in your desktop, Howard. So there's one thing I want to say. I don't know whether – so ESRP is, again, a synthetic workload. So I don't know if they're using real emails or using some sort of synthesized data for the email payload. It's synthetic data. They're just using the jet engine. Yeah, yeah. So it may not dedupe at all.
Starting point is 00:24:41 It may not compress at all. Or it may compress down to zero. I don't know. So you bring up some interesting points about synthetic workloads. So one of the tests that we see promoted pretty heavily with all flash arrays advocates for using an unreducible data set. And I think the reason for this is because it's the quickest way that you can try to fill the capacity on a data-reducing storage platform. Yeah. But the downside to that is –
Starting point is 00:25:11 If you don't have time to test multiple test sets, you look at the edges. Right. And irreducible is worst case. Yeah. Well, irreducible I think for a platform like ours is somewhere between a 5 to 6x load increase, and I think many miss that. So if you take our basic numbers, which is right around a 6 to 1 data reduction, and if I round these just for simple math on the podcast, if I'm providing a global average, meaning across my install base, a global average of 3 to 1 compression,
Starting point is 00:25:44 that means I'm actually getting a 3x amplification on the IO back end of the storage system. Absolutely. And when you introduce a non-reducible data set and then try to drive the system to max load, you get a false ceiling on your performance. Conversely, if I have a non-reducible data set and if my D to D duplication ratio is a global average of 2 to 1 that means I'm garbage collecting at a 2x
Starting point is 00:26:12 amplified rate because I'm now having to invalidate or garbage collect on blocks for cells that I've invalidated where normally I wouldn't invalidate them until the last reference has been removed. So it's interesting how systems work with test loads. Yeah.
Starting point is 00:26:30 The problem is that we don't have good benchmarks that use quasi-realistic data. You know, maybe there's a solution here for us. Well, I've been working on it. But, you know, the email thing is going to be ready in a couple of months. That's good. That's good. I'm sensing how it's going to be, send me one of your systems. Or send two.
Starting point is 00:26:54 What the hell? Why only one? Yeah, yeah, yeah. That's true. There's replication now. There's true. There's replication now. There's replication. But, you know, if you're using the common synthetic benchmarks, you know, basically you get to choose this data is going to reduce much better than the real world and the numbers are going to be meaningless. Or this data is going to reduce much worse than the real world and the number is going to be meaningless. And between those two meaningless numbers, i'd rather have the one that's
Starting point is 00:27:25 too low than the one that's too high you know so if i if i know i need a hundred thousand iops and that worst case comes out at 120 then i'm going to go yep got overhand i'm fine yeah right you know you could go the other way and do what i'm about to complain in a blog post I've been seeing from hybrid guys where people do reviews and they use test data smaller than the flash component of the system. And I caught one guy who said, to make it run as fast as possible, I did this. Since when is it a reviewer's job to make things run as fast as possible? Don't get me started on 2 million IOPS. Yeah, well, there's that. I've said too much.
Starting point is 00:28:12 Okay. All right, back to the product. So we've got five forms of flash reduction. What else is the thing, though, besides black storage and stuff? You mentioned replication. Yeah, so we have to start with a premise of disruptive economics and I think we've seen this, as I said, grow through
Starting point is 00:28:29 the market. I think there's other design details that are interesting. So today the system is a two-node modular storage architecture and I know people are going to be, boo, boo, I want scale out.
Starting point is 00:28:48 Did you hear either of us talk about it? Neither of us did that. I was referring to your listeners. Yeah, you're right. I think the challenge with that is sometimes we as a market, you know, enamor different technologies with an assumption that they're all the same. So my assumption when talking to a customer who says, I want scale-out, particularly for block storage, look, scale-out's a requirement for file storage. Let's not even debate that.
Starting point is 00:29:17 Scale-out in block storage usually provides you a means of simplifying the management of multiple systems. But what you really want is to be able to grow your system on demand non-disruptively. And when you look at the scale-up flash systems that are in the market today, they don't all do that. I might also want the efficiency of having one deduplication realm instead of two. That's true. That's true. But I think the basic premise, I think, when you talk to a decision maker remember someone who's not as technical is you know can i scale this system without reconfiguring my hosts or my environment and i don't know that that's a constant within the current
Starting point is 00:29:55 scale at all flash arrays however our system design clearly it isn't some there there are there are definitely products on the market that have a scale-out architecture but not a scale-out implementation of that architecture. Look, when you've got to migrate your data to either scale or to add new software updates and feature sets, some of us call that beta code. Okay. All right. I got you. Down. So what we have is we have a two-node or dual-controller architecture. It is built around a stateless controller framework, meaning the controller is simply CPU, memory, and I.O. ports. This architecture allows us to introduce some really new elements into your two-node storage controller
Starting point is 00:30:46 that I don't think have existed in the past. The first is an element around performance assurance. The controllers operate on the front end as a symmetric active-active array, all ports active, all IO coming through all the ports, simple multi-path and configuration management allows us to address the SSD on the back end as one global pool. And in this architecture, we're able to actually limit the resource consumption within the controllers to half of the actual physical CPU and memory. And what this allows us to do is to ensure the performance of normal operations remains the same when I enter into a maintenance mode operation or, heaven forbid, an unplanned outage.
Starting point is 00:31:33 And this framework we call always-on is a different way to look at resiliency. It's not about just having 5.9s or greater availability in terms of if I fail a controller or pull a drive. It's about never feeling the impact of a failed controller or a pulled drive. And this is really critical as you shift from disk-based arrays that served IOs or IAPs in the tens of thousands range into the hundreds of thousands range. But this is a problem we saw quite a bit. Right. Even with those disk drive systems where, you know, sure, it's active-active controllers, but to get all the features to work, you have to run all those CPUs at 70%. And then when something goes wrong, you limp along until the Maytag repair guy shows up. Yeah, two nodes running at 70% doesn't mean that one node can run at 140.
Starting point is 00:32:26 Nope. So it's interesting, because every now and then someone will say, well, I want that tunable. And it's like, you think you want a tunable, but you actually don't. The only way that you can assure this performance is by taking it out of the hands of those who are going to put load on the system and letting the system
Starting point is 00:32:41 enforce the limit. So that's the first element there. It's around ensuring just system state and performance during normal load. And I would highlight, again, for someone interested in a scale-up versus scale-up architecture, I don't think that steady state performance is seen on any of the all-flash scale-outs today. That's to the best of my knowledge, and I'm not here to start a FUD storm. It would be maybe best that you should inquire or ask your vendor what it looks like during failure operations. During an outage or something like that, yeah.
Starting point is 00:33:18 Yeah. Sorry. Let's pivot for a bit here. So the second thing that we have from this design is we're able to actually change. Wait, wait, wait. Change. Wait, Vaughn, Vaughn, Vaughn. You said stateless.
Starting point is 00:33:31 So what happens with cache when power dies and stuff like that? Ah. Are you really saying stateless? So flash controllers don't have cache in the way that you consider a disk controller. If you're going to say cache, then you can say we have 100% cache misses because we don't have cache. If you're saying flash and you're saying performance when data is from flash is good, then we have 100% flash hits. So what I mean by stateless is there's basically two tiers within a storage subsystem. There's the persistent tier, and if this was a disk system, you'd say that's disk storage.
Starting point is 00:34:09 With us, it's flash. And there's usually a non-persistent tier where you do your write buffering. That's called NVRAM. Our NVRAM components are not in the controllers themselves. They're actually in the shelves. Okay. So that's how we make it stateless for me. And the last time I looked were flash.
Starting point is 00:34:32 Today they are flash-based. So we have the persistent and the non-persistent tiers in the shelves. And again, as I said, from a read perspective, while there is system memory, that system memory is really more of a metadata table construct and being used by the system to run and operate, not used as a read cache as you have with the traditional storage system. as well. As you update the metadata, you potentially could be causing some serious problems. So that's also flushed out to this NV RAM into shelves. NV Flash, I should say, or just Flash. Just a Flash. Go ahead and go back to where you were going. Now we do mirror, just to be clear, the metadata constructs
Starting point is 00:35:24 in the metric in both systems. So that at failover time, we're not having to rebuild that construct or reread it from Flash. It's mirrored in memory between the two. Okay. But in terms of the classic node A cache gets lost and has to be rewarmed on node B, that construct doesn't exist. Yep. has to you know gets lost and has to be rewarmed on node B that that construct doesn't exist so now that we have a stateless controller architecture what our customers can can have again that's that's unique in market is they can they can increase their performance by simply upgrading a controller in place and it won't disrupt operations, including it won't disrupt performance.
Starting point is 00:36:06 And so we've had customers move through the FA300 series. They can go through all the products in the FA400 series, and we will take this architecture forward as we go into the next FlashArray series and further. This is an architecture that we're committed to. And basically, this allows a customer, start on two controllers of the previous generation, you know, move to the workload under one of those two controllers, take one out, introduce one controller of the new generation, shift the workload to the controller of the new generation, you know, subsequently remove the second older generation controller, drop
Starting point is 00:36:41 in the second new generation controller, and they're done. And that process is completed in about 15 minutes. Again, no change of configuration on the host, the network, or anything above the stack. And so if there is no performance pain and there is no IO disruption, did anything actually ever happen? Other than the fact you updated and potentially doubled the performance. We'll still schedule it for the weekend. Well, that would be because we're conservative kinds of people.
Starting point is 00:37:14 Well, that's an element that actually when we get into these benchmarking of systems, we're a big advocate of test the system how you would operate it, meaning go through completing upgrade processes. Now, controller hardware upgrades may not be something that you can do when you receive a system for proof of concept, but you can absolutely go through software upgrades and all the other failure scenarios and validate what can or can't be done during normal operations or should be scheduled for off hours. Yeah, I agree and think that folks in parts of the hyper-converged market are going to be unpleasantly surprised with the size of the impact of failures. Yeah, you bring up hyper-converged.
Starting point is 00:37:59 It's an interesting time to be in the storage market. Oh, yeah. The volume of innovation in this space, I think, has one resounding theme, which is the incumbent storage vendors and the incremental gains that they've put in their products over the course of the last five or six years just are not doing enough to meet the new paradigms of how data is served. They were designed for single hosts mapping directly to LUNST, and these compute clusters we have, whether they're virtualized or big data, et cetera, can really bury the
Starting point is 00:38:31 traditional array, and it's great to see this innovation. Now, the question I have will be, what's the lasting effect in the market? And my analogy here would be, many of us grew up with broadcast television where we had maybe five channels to choose from. I was in New York, we had seven. If I count public broadcasting in Detroit, I had seven as well. Because I had three from Detroit, one in Canada.
Starting point is 00:38:57 And you could watch the CBC. That's right. But anyways, my point was we went from a handful of channels to cable TV and now we have hundreds of channels. And market share, in terms of program market share, really became diluted. I wonder what's going to happen in the storage industry as we go from five major players from five years ago, and we go forward. Are we going to move to 10 major players, each with smaller market shares? Or are we actually going to see a change in the guard?
Starting point is 00:39:27 Yeah, I mean, don't you see a consolidation play here at some level where big companies buy small companies and incorporate some of the technology and, you know, I don't know. Some of the big companies aren't so good at that. Yeah, some of them aren't. Some of the big companies aren't so good at that. Yeah, some of them aren't. Some of them are. Including some Vaughn used to work for. And Howard pulls out the velvet hammer.
Starting point is 00:39:58 But it doesn't mean they're going to be poor for the rest of their life in this endeavor. No. And the interesting thing is actually that the marketplace – one way the marketplace could change is for Cisco and Lenovo to decide they want to look like HP and Dell and play for real in the storage business. I think Cisco kind of has to 18 months – 18 to 24 months from now to maintain their growth rates. Yeah. I think Cisco is vested. I know Invicta is currently in the middle of a stop ship, but they'll resolve that and get back to shipping product. Yeah, but they need a broader line. But they've got a lot of weight they can throw around, and I'd be surprised if they're not going to jump into the
Starting point is 00:40:49 hyper-converged space, whether it's a la EvoRail or that was Evo, not Evil. Yeah. Or some other instantiation of DAS-based storage architecture. They dipped their toe in the water with the SimpliVity deal.
Starting point is 00:41:08 Oh, that's correct. Thank you. Thank you, Harlan. They just did yesterday another generation of servers announcement, and the only form factor that made a lot of sense from the storage point of view is they did a 4U30 or 36 drive system that would be a good ZFS box. form factor or the high density form factor like the Super Micro Super Twin that Evo Rails is modeled on. So I think they're a year from being serious even about that.
Starting point is 00:41:58 In no small part because they get a boost in the market from being the servers every storage vendor uses in their reference design because they're the only server vendor that doesn't have a storage division. Yeah, yeah, yeah, yeah. Yeah, they've made a lot of inroads with a lot of storage partners, and they have a significant presence in market. Now, what I think could be more interesting for Cisco besides storage is can they continue to grow within the Blade ecosystem as it seems like Blades are starting to become less desirable? Yeah.
Starting point is 00:42:34 Yeah, no, I think the Blade era is coming to an end. Yeah, because their market growth is strong in Blade. Because you can't do hyper-converged on Blade. Yeah, and their market growth was strong in Blade, but not in the overall server market. Right. The other interesting thing they announced the other day is a modular compute server. I didn't see that. It's got 16 Xeon E3 servers in a 2U frame.
Starting point is 00:43:04 Nice. There's no I.O. at all, just shared 40 gig Ethernet amongst them all. But it's good for more moonshot than to run VMware on. Right. For compute applications that you're going to shard across a million servers. All right, all right. We need to get back to Pure here. Okay.
Starting point is 00:43:30 Was there other stuff you wanted to talk about? The product? Yeah, so we were talking about the architecture when we kind of went off on another little jaunt. I have actually said many times that I think the fact that at first glance a pure system looks accessible to a storage guy was brilliant in terms of speed to market acceptance. Yeah, yeah. You didn't have to teach people about 10,000 other pieces of technology and why they wanted them. Yeah, I believe there's a formula to success, which is if you can make the burden to adoption low, and what are we offering? We're offering a storage device that provisions externally LUNs. You connect to them, you move your data on them, and then you measure results.
Starting point is 00:44:20 So if the burden to adoption is low and the results are off the chart, you have the formula for fantastic success. And it takes much more than just the storage device itself or disruptive economics and data reduction technologies. You've got to have a better experience through the software licensing model, the availability of the platform. And to your point, what's my operational burden? No knob turning is sometimes a little too much for some people to grab. You sit there and say, there's no RAID layout,
Starting point is 00:44:52 there's no root volume, there's no aggregates, there's no cache tuning, there's no block alignment or assignment. Yes, there's some of the data reductions in line and some is post-process. The problem is you're talking to guys who think they have a job because they're the keepers of arcane knowledge.
Starting point is 00:45:09 And now you're saying that there is no arcane knowledge, and they're wondering why they're going to keep their phony baloney jobs. And the answer is that the really arcane knowledge isn't how to tune the knobs. It's how to know what you need. Yeah, I think in terms of the role, kind of like the role of the backup admin had changed, right, with the introduction of disk into the tape world, where disk became the backup medium and tape became the archive medium. I think those who are in charge with maintaining storage, if they can start to think of their role as being the ambassadors of data and the fact that your storage systems are going to radically change this. What we're seeing today with Flash is just the tip of the spear because structured data works very well on something like a pure storage device or the other all-Flash arrays I've mentioned. But the bulk of the net new data that's being created today and into the future will be
Starting point is 00:46:07 unstructured. And that's going to probably be a better target for maybe tier two platforms. And, you know, you're going to have to learn to make a choice on those designs as well. Do I just throw white boxes at them and think that's the cheap mechanism based on open social software, and there's folks that are in that camp? Or do you say that those systems, the shared nothing architectures, have an inherent overhead of 3x footprint that works against the constraints of the data center that we mentioned earlier in the podcast, And I'm more in that camp.
Starting point is 00:46:47 So lots to learn. It just means your knowledge base as a storage admin has to change. And by simplifying what you've done historically, that's a good move. Yeah. Okay. So we're over 40 minutes here. Any last questions, Howard? No, I'm pretty good.
Starting point is 00:47:06 Okay. Well, this has been great. Vaughn, do you have any last questions for the graybeards here? Yeah, not a question, more of a comment. And for those on the podcast that were unaware,
Starting point is 00:47:20 I participated in a panel with Howard and a number of other representatives from some storage startups at VMworld. And I closed that session with the same way I would close this one, which is while we at Pure Storage believe that we're making the best all-Flash array platform, I would advocate for you to check out Flash from whichever vendor or platform you think best services your needs. Because if you can get the economic equation to that of disk, you will have a better experience in Flash than you've ever had on disk. And so, not a commercial for us. It's more of an endorsement for this medium and the adoption of it.
Starting point is 00:47:58 Well, I think that brings it all home. This has been great. Thank you, Vaughn, for being on our call. Hey, thanks, guys. This is great. And next month, we will talk to another startup storage technology person. Any questions you have, let us know. Thanks. That's it for now. Bye, Howard. Bye, Ray. Thanks again, Vaughn. See you guys. Until next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.