Podcast Archive - StorageReview.com - Podcast #118: Going Deep on Microsoft Azure Arc, Cloud, Managed SQL, AVD, HCI and More!

Episode Date: March 28, 2023

Brian invited Ernie Costa to join him for this podcast. Ernie is a Team… The post Podcast #118: Going Deep on Microsoft Azure Arc, Cloud, Managed SQL, AVD, HCI and More! appeared first on Storag...eReview.com.

Transcript
Discussion (0)
Starting point is 00:00:00 . Everyone welcome to the podcast. I'm Brian Beeler and today we've got an interesting conversation. We've got an MVP in our midst. So we'll see what that's all about and learn quite a bit more about the Microsoft Cloud infrastructure, what's going on there, what's going on via Azure Stack HCI and some of
Starting point is 00:00:22 the services and other things Microsoft's offering there around SQL Server and other database offerings. Ernie, thanks for coming in and doing this today. No problem. Super excited. This is going to be fun. Yeah, so you're an MVP. In the Microsoft world, everybody knows what that means. If you're not knee-deep into Microsoft, what's an MVP? So I'm relatively new to this as well. So you don not knee deep into Microsoft, what's an MVP? So I'm relatively new to this as well. I've been an MVP. So you don't know the answer to that? Yeah, no, it's kind of like, I don't know, I gotta go Google that.
Starting point is 00:00:50 Great, great start. Yeah, right? No, so I've been an MVP for about a year now. Renewal is actually coming up. MVPs are resources that Microsoft has identified folks and individuals that have contributed to the tech community over the years that are actively engaged with customers or even just folks that maybe are running their own home lab in their basement or something. They don't necessarily need to be a professional and tech being their profession. I just happen to work for an ISV. But the two sometimes overlap, but they don't necessarily need to. You know, it's funny because the MVP program has been around forever and it's, Microsoft used to use it in a way to incentivize the use of their products amongst highly active and engaged people. Now it's probably anyone with a couple, you know, tens of thousands of followers on TikTok or
Starting point is 00:01:42 whatever it is, right, to help do the evangelization of Microsoft products. But way back when I used to be an MVP for this product called Spot. I don't know if you've ever even heard of this, but they used to put, Microsoft put transmitters in radio station tower infrastructure and it didn't do much, but it would broadcast out over the radio signals, things like weather and news. And it didn't do much, but it would broadcast out over the radio signals, things like weather and news. And I think you could even set up personalized alerts. I mean, this is before everyone had a cell phone in their pocket or maybe at a similar time. But yes, a very short-lived technology that is nowhere near relevant to much of anything today. But it's funny, that stuff, it kind of lingers around with I think like HD FM radio like you see like remnants of it still there and at
Starting point is 00:02:30 least in the U.S. so it's kind of cool tech that that you know it it moved on to something else. Yeah but you know to be fair like the radio kind of always works so while it may not cover everybody in in all places and you know if you're on the back side of always works so while it may not cover everybody in all places and you know if you're on the back side of a hill but radio is a technology is still pretty neat and highly capable for what it does right yeah so we didn't join this uh this pod together to talk uh radio and spot technology uh one other disclosure though so your your day job though, while you're an MVP and wear a cape at night, your day job is with Commvault. Yeah. Yep. So I work for Commvault. If you're familiar with our backup software, I do not, I'm not a software engineer by trade. I work for our
Starting point is 00:03:16 internal IT department. So I am using the technologies a lot that we're going to be talking about internally. It's not stuff that necessarily we're selling to our customers or telling our customers to use. It's more us as a medium-sized enterprise, what we are using internally there. Right. Well, that makes sense. I just want to make sure if you start slinging mud at like Veeam or Haiku or Cerdo that just everyone knows.
Starting point is 00:03:41 Ernie, go off the rails and start talking trash on the backup provider. We have, in the MVP community, the good thing is there's a lot of folks that do use other software and we actively are engaged with each other about, you know, kind of common things we may see acting up between or even with the various different pieces of software that are out there that do kind of the same thing. So it's not too violent.
Starting point is 00:04:05 It's not Yankees and Red Sox or anything like that. All right. So you're in the Microsoft world. We prefaced this conversation with that. And you're in the IT organization at Commvault. So that's all good background. What is your exposure then to Microsoft products? What are you using on-prem cloud in terms of hypervisor software? Just what's your consumption look like? We are overwhelmingly a majority Microsoft shop. We tend to have a very large presence in Azure for production workloads as well as some dev test type of things. But we also have a traditional on-prem. We have our own data center in our headquarters.
Starting point is 00:04:55 We have a couple of co-locations and remote offices throughout the world. So we still have a physical presence on-prem as I'm sure a lot of enterprises do. So hybrid connectivity since like 2018, 2019, since that buzzword started getting thrown out, hybrid cloud, we've been kind of right there, nose deep in it, getting our hands dirty with it in some capacity. So again, majority Microsoft, Hyper-V, and most recently Azure Stack HCI as our production hypervisor workload. That's interesting because still the predominant hypervisor continues to be VMware ESXi.
Starting point is 00:05:33 I'm curious, why Hyper-V? What has that done for you guys? Had you been on VMware before? What's the history there? Without going too much into it, the short answer, as I'm sure everyone probably knows, is licensing and pricing. So it's expensive to have a large VMware footprint. If you're committed to it and your infrastructure and automation is built around it from day one, then you may have a wrangle on it. But especially with the recent acquisition, costs go up. Maybe or maybe not, you're happy with support that may have been coming along with it. Again, we are a Microsoft Gold partner. It just seemed like a natural fit for the workloads that we were running. And this is decisions that were made before I was involved. But it just was a natural
Starting point is 00:06:16 thing. If we're going to be running a lot of Windows server workloads, and we're already paying for the, at the the time Windows Data Center licensing, why wouldn't we just use the hypervisor built into it? So that's kind of the short, sweet answer. That makes sense. And so one of the things you also mentioned there was Azure Stack HCI. So I want to understand that a little bit too, because where we're going to go is going to talk a little bit about Azure Stack HCI, Azure Cloud, and then some of the services in between.
Starting point is 00:06:47 We've talked a lot about Azure Stack HCI. We've written a lot about it. We've done a couple reviews with DataOn on some of their all-flash and hybrid configurations. I've had Cosmos on the podcast and talked about Azure Stack HCI quite a bit. And the one thing that he was excited about last time we did the pod was that in the old days, Azure Stack HCI was kind of like a feature of Windows Server, like a checkbox of turn this server from a server into an HCI node basically, and join up with some others and off you go. But it feels like when he and I talked,
Starting point is 00:07:23 it's probably been six months, and I'll try to remember to link it in the description here for those that want to check that out. But it really feels like that at that point, the intent was to really mature the product and make it its own independent stack. I'm curious how deep you are on the on-prem HCI product and whether or not that looks real to your eyes. To summarize it, we are pot committed to Azure Stack HCI at this point
Starting point is 00:07:53 from our internal IT perspective. Like I said, we've been running, do we have a VMware footprint? Of course. Is it large? Absolutely not. Have we used legacy Windows Server in the past? Yes. Do we still have some? Yes. Do we use NetApp and other SAN-based storage? Of course we do.
Starting point is 00:08:15 But there just seems to be this ease of management that our IT department has kind of fallen in love with when it comes to the software-defined defined storage stuff that you get with, uh, storage spaces direct. And then obviously the evolution of the OS into Azure stack HCI. Right. Um, there's, I mean, I can sit here and go on for hours about the, the value add in the bonuses of HCI, you know, your billing, your support experience is, is,
Starting point is 00:08:44 is significantly better than the traditional enterprise support experience that you may be getting with traditional server SKU. I think I've done a few videos detailing like top five reasons why you should move over if you're interested. From a financial perspective, again, if you're already paying for data center licensing, and you have software assurance, it kind of is free, like nothing's free, but it's you're already paying for data center licensing and you have software assurance, it kind of is free. Nothing's free, but you've already paid for it, you might as well look into it. The update cadence is significantly faster.
Starting point is 00:09:13 New features are coming to GA significantly faster than with long-term support branch. So all those things just kind of fit right into our vision. So it's kind of a no-brainer. you're hybrid though, too, it seems to me, and I'd be interested in your perspective that Microsoft with the Azure Stack HCI product, the on-prem bit is really trying hard to take the cloud Azure features and functionality and drive those into the on-prem product. And I think we're a long way from feature parity between the two, but it feels like with AKS, with virtual desktop, with some other stuff, they're trying to really push that down from what they've learned on cloud to our governance and our proper security structure and all that happy stuff that you have to worry about in the cloud. How do we wrangle it out on-prem? Because even though there may not traditionally have been a cost associated to it, you still have to worry about security policy and just kind of your posture on-prem of not letting it be the wild west like it may have been.
Starting point is 00:10:25 And a lot of the Azure Arc-related things that you mentioned have simplified that dramatically for us. Again, automation is kind of the – our fiscal year is about to start, and that's like our North Star. Like, how can we handle all this stuff? And it's kind of, oh, Azure Arc, like one line right there. We can do it all with Azure Arc. So we're excited to be transitioning to that as kind of our guiding thing. Well, you said I mentioned it, but I never said the Arc word. Oh, I'm sorry. Jumped the gun. Mr. MVP is, like I said, very deep in the Microsoft technologies.
Starting point is 00:11:06 It would be my theory, my supposition, that most people don't really understand what ARK is or the nuance there. So I know you started to go down that road. But take half a step back for me and just do a hit on, from your view as a practitioner, this will be really interesting, not as a marketer, right? In your view, what is ARK? What does it encompass? So if you were to just ask me as like a lay person explaining it to my father, who's also a lay person when it comes to tech, I would turn around and be like, it's a branding term. It's a collection, a suite, a bunch of different technologies that Microsoft has enabled to be cloud connected. So let's say you have SQL servers on-prem or hosted in AWS even, and you want to kind of control them from a single pane of glass, or you want to just monitor usage, you can leverage Azure Arc in some capacity. If you are large in the Kubernetes ecosystem and
Starting point is 00:12:07 you have maybe, again, AWS EKS stack, so that's Elastic Kubernetes Service, or AKS in the cloud, that's Azure Kubernetes Service, or even you're just running your own Kubernetes cluster on a Raspberry Pi in your basement, you can connect that through Azure Arc to your Azure subscription and tenant, where then you can, again, manage it through a single pane of glass. You can leverage command line. You can use it as a jump off point for automation, a host of other things. And those are just two of the many Azure Arc connected services that exist right now. Well, let's pick one.
Starting point is 00:12:43 Let's pick one because like SQL Server, I'm sure you guys do a ton of SQL. It's, you know, one of the most ubiquitous databases out there. When you think about how SQL is made available or visibility for SQL or manageability is made available through Azure Arc. It's a bit of a spectrum, though, isn't it? Because some you're talking about some of the visibility into what's going on in your entire ecosystem. But then there's this whole other managed services bit that I want to get into and understand too. So kind of break that down and let's just use SQL as a baseline app that most people understand. Yeah, so when you're thinking about SQL and if you were to ask someone on-prem,
Starting point is 00:13:27 hey, I need a bunch of SQL servers spun up for something, whether it's a production workload or maybe some developers are testing some software and they need just a boatload of SQL to test it against. Traditionally, if you're familiar, it's a somewhat manual process of installing SQL Server. If you want to worry about high availability, you need multiple servers to spin that up. You need the right storage and you need to worry about fault tolerance and all of that. So there's a lot of unique skill sets and fault domains you need to worry about, just to even set it up and sprawl it. With something like SQL Server managed instances
Starting point is 00:14:15 that's deployed on-prem and managed through Azure Arc, you're leveraging the, I'm going to start throwing out a lot of words here, so let's try to keep it simple as possible. You're going to be leveraging a small little virtual appliance
Starting point is 00:14:33 that lives on your hardware on-prem, and that appliance is going to be communicating up to Azure, and when you say, I need a highly available SQL deployment, it will phone home to your hardware on-prem and it will spin out a couple containers of SQL Server. So ultimately, this is running a Linux version of SQL because that's a new thing now that you can do with SQL
Starting point is 00:15:03 Server. You can have a Linux build of it. It's running inside a container image. If you're familiar with containers and Docker, it's running on a similar engine to that. And it is scaled out using Kubernetes that is also managed for you across all of those pieces of the hardware that live on-prem for high availability. And your storage, which is already clustered and configured through the magic of hyper v or if you're using vmware you can also use that um it is able to um leverage the existing high availability nature of that and now you just have these pods that are pods containers that are running across all this hardware uh certain vCPU size, certain memory size that are based on some parameters they may pass them when you said to go deploy it.
Starting point is 00:15:51 And let's say there's some kind of catastrophic failure on one of those containers. Well, the magic of Kubernetes will spin up a new one for you and your data stays resilient. So again, I'm throwing a lot of words out there, but at the end of the day, the whole architecture is built around you not needing to worry about that high availability because it's all handled for you at deployment time. So as long as your prereqs are met, which is give me a hypervisor to put this on, the product itself handles the actual management and lifecycle of it. Well, it's interesting because I think it matters who you're talking to, because if you're talking to any app developers,
Starting point is 00:16:27 what you just described, first of all, they'll probably never see it. They just know SQL's working and they don't really care. And you say things like containers and they nod along because yeah, we get that sort of lifestyle. The traditional application managers though, when they hear SQL on Linux in a container, probably had some sort of blood pressure rise of 20 to 30 points. Yeah.
Starting point is 00:16:55 What have you guys found as you've been working on this in terms of acceptance? And it's just kind of weird, right? It is. acceptance and it's just kind of weird right to think about doing it that way it what what have you seen in terms of your reaction from practitioners and their willingness to to say yeah maybe i don't want to do the legwork of installing and updating sql maybe it would be nice just to you know kind of click a button and go so that that's that's a major one. The actual not having to worry about patching an operating system and then a build of SQL that also lives on that operating system is huge because I can tell you right now we have some large on-prem traditional SQL server availability group
Starting point is 00:17:36 deployments and despite them being virtualized and us having enough high availability and resiliency to reboot one at a time and update one at a time, it's a headache. It is a very orchestrated song and dance just to get us through that patching cycle. And it's not that it's difficult, but it's time consuming. The beauty of containerization is you have something called a base image with which contains those binaries of SQL Server. And those don't necessarily have the persistent data of the database itself in them. So when you're just updating, you're just saying, hey, spin down this image, spin up a new image with the new binaries, and just attach the data back to it. And it's not instantaneous, but it's pretty darn near instantaneous.
Starting point is 00:18:26 So the patching and update cycle is reduced down to minutes instead of hours in some cases, or days depending on how many machines you're dealing with. You also mentioned the kind of obfuscation of the back end. One thing I've been trying to get through
Starting point is 00:18:45 to our developers for years now is, and this is a term that's very popular in the IT world, stop treating your servers like pets, treat them like cattle. And what I mean by that is it's a SQL server. Who cares if the operating system, like unless you're doing very, very customized customizations and tweaks and stuff on the OS, which I don't know why you would be, if something bad happens to that SQL
Starting point is 00:19:13 server or that OS, it doesn't matter. As long as the data integrity of SQL itself, the databases, the MDF and log files are still good, that should be all you really care about at the end of the day. So again, with this containerization portion, yeah, the traditional admin may turn around and kind of crawl in their skin a bit about thinking, oh my goodness, I need to be familiar with Linux now. I need to be familiar with Kubernetes and containers and how do I troubleshoot things?
Starting point is 00:19:41 You kind of don't. Again, if the data is good, kill a pod if it's acting up, and it'll just spawn a new one for you. And assuming the data is good and copacetic, it'll mount it and you'll be back up and running. And if for whatever reason that container, that specific one of N amount of containers for your high availability is not working. The natural resiliency of Kubernetes where there's health checks built in to say, hey, can I even listen on port 1433? All those things are kind of handled and it removes things from the load balance equation so that you're not hitting a bad node, for example. So those are the two biggest things for us. It was getting to the point where we don't treat our servers like pets. We treat them like cattle,
Starting point is 00:20:33 and we eliminate them when it's time to. Turn them into juicy steaks. You know, it's funny, I'm laughing because you said don't treat your servers like pets. When we started, not StorageView, but the company before, we hosted all of our own stuff. We did have some databases. We had all the stuff that growing small businesses have. And we absolutely named every single one of our servers, which is absolutely hilarious. We named them after famous pigs. Okay. Yeah. So we had Babe and we had a big storage server that was Porky and we had, I don't know, three or four other ones. The fact that I still remember the first couple is surprising. There is sometimes a fond emotional attachment to hardware. I've been in a thousand labs, and I'm sure you've seen the same, where the bezels will be there, and then there'll be the yellow tape, like the little printout tape that'll have a server ID,
Starting point is 00:21:37 but oftentimes it's a human name or some sci-fi thing of Borg 33 or whatever. Big ones for us were definitely like Greek gods. Yeah, I'm sure that's one. Transformer names. And then I think in college, my college lab, I was naming them off of budget beers. So there was like Bush, Natty, and yeah, oh yeah. Red Dog, just like ridiculous things.
Starting point is 00:22:26 But yeah, but again, we're talking, you know, when you're talking 15, 20 years ago that those were that was the norm. And I don't necessarily mean naming your computers after beer, but but the the concept of like. just treating them like they're your babies and keeping them up and running was a full-time job. And at the scale that technology has grown at now, you just can't. It's impossible. You're right. I mean, it's different if I'm sitting here with seven servers in a baby half rack in my SMB. But if I'm a serious business, you don't even have to be enterprise scale.
Starting point is 00:22:44 Sure. A decent-sized operation, If I'm a serious business, you don't even have to be enterprise scale, just a decent size operation. Having somebody go in and manually run all that stuff on Tuesday is tedious and it's expensive. And depending on how big your IT org is, and I think that's really the challenge for SMBs is the size of the IT org. You might only have one person that's skilled in doing that. And if they're not there or something happens or they quit or whatever,
Starting point is 00:23:13 you've got a problem. So I do like this notion of as-a-service consumption from an OPEX standpoint of, look, it just happens. And so long as I can get support, if something goes wrong and, and, and that's one of the things you were talking about before, then I, then I feel okay about that. In terms of what you're doing with SQL, managed SQL, or I actually, I'm assuming, are you guys using this? We, we have, we not in product. Well, it depends on what your term is production but no
Starting point is 00:23:46 we're not using it from a we are using it at our workplace for our developers and our engineers to rapidly spin up and spin down kind of like test environments for the software that we write um traditionally that would have been a vm VM with a sysprep image of Windows and SQL Server there, and then there were manual steps that needed to occur. But from soup to nuts, start to finish, that process was taking 10-15 minutes, and now it's
Starting point is 00:24:16 down to 10-15 seconds to make a SQL Server on demand using ARC and the whole suite of tools. But no, we are not running production workloads on it from a customer-facing perspective, only because of the way what we sell doesn't fit into that category.
Starting point is 00:24:38 Right. Okay. But you still see the operational benefits, even on a smaller scale for your app dev teams to be able to get in there to quickly spin things up and theoretically tear them down as well. Absolutely. So for those teams, if they were to consume this through your on-prem
Starting point is 00:24:57 Azure Stack HCI or Azure Cloud, how similar does it feel and where might there be a little bit of a disconnect, if any, maybe not? I'm not sure from a practical standpoint. So from a traditional, traditional being Azure cloud, public cloud, if you're not manipulating this through like Azure command line or Azure PowerShell, you're just clicking, clicking through the Azure portal and saying, hey, give me a SQL managed instance with this amount of vCPUs, and you've got a little slider, and how much memory can be allocated to it, and the storage account or the disk that's attached to it. A lot of that kind of bleeds over into the on-prem managed side of things. You have some similar UI portal nuances, but there's some key differences. And it's mostly, again, about the access you have to the
Starting point is 00:25:57 back-end infrastructure. So meaning you're running this on hardware that lives in your data center. It can technically live in Azure on a VM as well. I don't know why you would do that other than just to kind of like test it. But you are managing and you are responsible for the physical hardware and the networking there as well. And while I did say earlier that, you know, oh, your pods and your deployments, you know, if they're healthy, you're good, and if they're not healthy, you just delete them and you spent there, there obviously can be
Starting point is 00:26:28 like a hardware fail. Like we've if we've been in a data center, you've had a failed rate controller happen once or you've had a disk die on you, or you've had, you know, backplane issues, those types of things can still happen. And you as the customer are responsible for for administrating them so there is some back end administrative tools that you have available that you wouldn't have had available in the public cloud um you also can just directly manipulate the uh again i mentioned earlier this is all orchestrated for kubernetes so if you're familiar with Kubernetes, you can also kind of dig into it through the Kubernetes controller. It's called kubectl. It's the command line tool to manage Kubernetes. You can also get into it that way and kind of see what's going on and troubleshoot and scale things out manually if you wanted to. Again, you basically have more things exposed to you on-prem
Starting point is 00:27:27 versus in the cloud where a lot of it is, again, obfuscated behind a pane of glass. Okay. But you still, through Arc then, if you're running in your hybrid environment like this, you're still getting the visibility into all of these workloads regardless of where they are, right? Correct.
Starting point is 00:27:44 Yeah, you're going to see databases and performance stats and metrics and all of that, and you can control it directly through ARM, ARM being Azure Resource Manager. So if you had automation and templates made to say like, hey, Azure, here's a template file with these 20 SQL servers that you've made with these settings and these names and this type of performance,
Starting point is 00:28:12 these metrics, CPU, RAM, you can shove that into Azure ARM and it'll spit it out on-prem for you, again, without you having to actually be at your keyboard in a cluster typing and saying do this and a lot of that a lot of that is analogous to what you would have already been doing with azure um with with traditional azure cloud managed sql they've they've kind of like unified the terminologies yeah so when you think about managed sq, then we talked about deployments fast and easy.
Starting point is 00:28:46 It's resilient, sort of inherent in Kubernetes. Patch management kind of goes away and is automated and part of that service. Is there anything else at a high level that's really important to call out, either operationally or just you know just from a from an overall benefit standpoint so there's definitely benefits of running it on your own hardware because you get to control like the the nuance of performance um right in the cloud you tend to the traditional cloud you tend to pay a lot for uh premium or ultra SSD, for example, or a storage account that may allow you to have a certain amount of IO or even like inter-site bandwidth issues.
Starting point is 00:29:33 If it's on-prem and your workload that's touching the SQL server is also on-prem, you obviously get the benefit of it being hyper quick in terms of latency, because you're probably in the same data center. So you don't need to worry about the internet or Azure's backbone being a bottleneck. You also get to control the storage that it's running on. And that's probably something that you guys see every day. If you know that you're running something
Starting point is 00:29:59 that's like super read intensive versus super write intensive. Well, you can, if your hardware on-prem can be customized from your OEM of choice, you can say, I want this specific type of NVMe or this specific type of SSD, or maybe I don't even care about the IO or latency in the sub-microsecond range. I care more about the capacity. I have a petabyte of SQL data that is kind of right once and red never,
Starting point is 00:30:28 but it just needs to be there. So you can throw a bunch of spinning rust at your cluster and well, it's there. You just get a lot more customization for your workloads. And that's not something that you can do very easily in Azure. And even if you can,
Starting point is 00:30:43 it's going to be significantly more expensive because you have to pay for that storage in Azure on a monthly basis. That's always the rub right is you get it fast and you get it easy in the cloud but as soon as you as you say want to drip into some of the more premium offerings around storage or connectivity I mean the cost starts to get out of control. But during COVID, when we had all these accessibility and supply chain issues, the cloud was still there and it was doing a pretty good job. And I do hear a lot about trying to take some of those expensive workloads and repatriate them and put them together. For me though, Azure Stack HCI has been such an anomaly because it's so much more cost effective
Starting point is 00:31:30 than a lot of the other HCI solutions. And it's really fast. It's faster than vSAN. Yeah. Now we haven't tested vSAN's ESA, but up to that point, everything we saw, Azure Stack HCI was always faster and by a wide margin. I do think, especially for smaller orgs, SMBs, it's a great place to consider an investment if you're a Microsoft shopper, can become one.
Starting point is 00:31:56 If you have an Azure subscription available to you, it is a no-brainer to just go download the trial, install the ISO on some kit that you have lying around and just give it a spin. You basically get it for free for, I don't remember if it's 60 or 90 days at this point. While raw IO is so important to a lot of people, it's a great metric. I, in my world have found that the amount of IOPS, if I'm getting 2 million, 6 million, 12 million IOPS, that to me isn't necessarily the most important thing. It's more the latency that I can, or the lack of latency, I should say,
Starting point is 00:32:38 that my workload is possibly generating and being affected by. But again, it's also that support structure you get with the product. I can remember back in the server 2016 and 2019 days, just to just keep apples to apples here, so to speak with Microsoft. When Storage Spaces Direct first came out, there were a lot of issues, there were a lot of problems. I had to open up enterprise support tickets all the time. And it got to a point where the support experience wasn't great for one reason or another, whether I was getting someone that may not have been technically skilled to answer our unique questions, or we just had some weird bug that we ran into and we just needed to get the product group involved and we weren't getting that channel.
Starting point is 00:33:23 The difference is with Azure Stack HCI, when you have a problem, you are opening a ticket through the Azure portal. And again, it's this unified experience where it just kind of comes free with the price you're paying for Azure Stack HCI on the monthly basis. And they have a unique set of engineers that are just tailored specifically to this. You're not going to be getting third-party contractors that maybe they're outsourcing or working with on the side. It has been a night and day shift. And again, in our world where it's critical, we need to have that support that can actually hold us up in problem times. Well, that's a hard thing to articulate too because Microsoft will say in a press release
Starting point is 00:34:09 or in some sort of scope document around Azure Stack HCI, we have this stuff, but until you need it, until you use it, most people when they buy these things don't spend a lot of time creating provisional tickets just to see what the experience is like. So the story that you just told is one that needs to be told more probably because, again, that's a tremendous part of the value prop of the ongoing OPEX benefit of doing any of these things either as a service and sort of even on-prem Azure Stack HCI is sort of service-y, I think, in the way they bill it as sort of a subscription, right? So you're paying
Starting point is 00:34:52 for the hardware, but then you've got the licensing for the software that goes with it. But that needs, if you're going to deploy that model, that needs to be supported and it has to be a really great experience for customers or they'll be gone on the next week absolutely yeah i can't tell you the like if it's 365 days in a year we have a production workload running and you know five days of the year we have an outage let's say which is a lot um the aggravation and stress and the receding hairline that that is a result of all that stress causes. And if I just had a better experience, how like I could. It's one thing outages happen, but it's another thing when you're dealing with with with something that just isn't getting the movement that you expect it to.
Starting point is 00:35:37 And I just have in the past two and a half years that we've been running HCI as our primary production workload hypervisor, just the whole suite of everything. I have yet to run into that experience. I've opened up dozens and dozens of tickets, and every one of them is closed within days, like full problem resolved. Again, I'm not just blowing smoke. It has been night and day.
Starting point is 00:36:01 Well, that's an important consideration. We talked a lot about SQL. I think I might have managed virtual desktops and AKS as we sort of drove through the notion of Microsoft really wanting to take these cloud services and push them down to Azure Stack HCI. Do you have any hands-on experience with either, or are there any other services you guys are using that you're excited about that can transition between your hybrid cloud scenario? Yeah, more so than even the SQL solutions that they have. AVD on HCI, which is Azure Virtual Desktop on HCI, is the next big thing that's coming down the pike. GA has not been announced yet. It's been public preview for well over a year now.
Starting point is 00:36:49 For a while, yeah. Yeah, you can think of it as kind of the competitor to like VMware Horizon or Xen, Citrix's product. If you're familiar with AVD in traditional Azure, it's basically spin up some VMs on demand in Azure and users can RDP to them through the remote desktop client app, the new one, not the one that's built into Windows.
Starting point is 00:37:13 And you're able to have pooled multi-session machines or personal one-to-one assigned machines. So that's now available on-prem through AVD on HCI. We are very heavily interested in this, again, with remote work and kind of avoiding large CapEx purchases for like laptop recycle, hardware recycle refreshes, things like that. And then also being concerned about data governance and another one that's thrown around. Yeah, security. But then another one that's thrown around is data gravity.
Starting point is 00:37:50 And I'll explain that one in a minute. That was kind of the reason why we're like, hey, this is where we got to focus our energies on as the next thing. So it's still in public preview, ABD on HDI. We're hoping to get some more information in the next coming weeks or months on general availability and pricing and kind of what that's going to look like and then the roadmap for all the features. But we're super excited about it. Well, I mean, any more with the the thin clients are so nice now these days. Yeah, the you know, bring your own device and having just an RDP session, as you said, to get into your work machine. The technology to push pixels over the line has become so good.
Starting point is 00:38:35 Yeah, we do. We do a ton of thin client reviews and user computing where we're looking at at these you know workloads specifically and yeah i mean you you may not necessarily you know want to edit 4k ak video on your general purpose machine but you can amp these things up of course absolutely things like gpus and now if you've got a a big distributed workforce well rather than putting a you putting a several thousand dollar GPU in every system, if I had some in the data center and we just managed access, I mean, that goes a long way to making teams more efficient. And around the security bit, the pushing pixels is a lot safer, all that data stays in the data center, and it makes it a lot safer that all that data center or all the data stays in the data center yeah and it makes it a lot harder to to leave a laptop behind in a in a cab or an uber i guess
Starting point is 00:39:31 or whatever and uh for international teams you might be worried about uh corporate espionage things like that i mean those are real problems data exfiltration is the number one concern that keeps me up at night and i tell that to my security guys. Yeah, you said it best. Pushing pixels is a lot more secure than even a VPN connection. That may be encrypted, but if John Doe is using his work laptop, or I should say his personal laptop, and he installed VPN client on there and forgot to disconnect. And then his kid goes on and decides to start doing whatever on there. Lord knows what could end up happening between that tunnel from point A to
Starting point is 00:40:13 point B, you know, whether it's crypto where and you can have all the endpoint protection you want and firewall rules you want. But if you open up that door, they can get in. And the beauty of remote desktop is it's literally just like pixels and audio and video it's it's it's not really a big concern or i should say it's less i just love having i just love having the stuff in the data center so that you can back it up i mean that's well that's that too so you you mentioned um a dispersed geographical workforce and we have a
Starting point is 00:40:41 lot of that here we have a large uh we have a large workforce in australia we have a lot of that here. We have a large workforce in Australia, we have a large workforce in India, and our main data center is in New Jersey. So when these folks need to manipulate large files, tens of gigabytes in size, doing that over a VPN connection is basically like, it's not impossible, but it's pretty darn improbable to have a functioning, you know, worklife balance. You're just going to be sitting there forever. So giving them jump boxes, if you will, that live in our data center, and this is that term data gravity, the data is the big thing that necessitates performance here. So having their VMs local to that data allows them to manipulate that data significantly faster
Starting point is 00:41:28 at the cost of some latency with the RDP session. But that's just screen redraws and mouse clicks. That's a lot faster than moving 10 gigs of data over the wire and who knows how fast that pipe is to even get it across the water. So that was the other big reason. You said backups obviously are great. It's local or data center. But just the user experience is significantly better. Yeah. So you're doing virtual desktop.
Starting point is 00:41:58 Are you doing some on-prem now? Are you doing that in Azure? We are heavily, heavily testing it on-prem. We are currently leveraging Azure public AVD, like traditional AVD. That's kind of like a fail over type of scenario. So if the 14 internet connections to our data center went down and no one could get in, yeah, okay. We spin up some VMs in the cloud and they connect to those. But yes, we have a couple hundred VMs that are being used individually by users as personal session hosts. And then we also have a number of multi-session Windows 10 and Windows 11 VMs that live on-prem. And that's kind of where this is going to be big for people. Getting access
Starting point is 00:42:46 to Windows 10, Windows Client SKU multi-session on-prem and not have to be in the cloud. And you can load up like 10, 15, 20 people onto a single VM. Yeah, it's got more cores than maybe your normal one-to-one, but you can densely pack that thing up. And again, if you got GPU partitioning enabled, man, oh man, there are some potentials for big cost savings there. Yeah. I mean, that's the trick, right? Is these, the evolution of these products tends to be, you know, Gen 1 is pretty baseline, task worker friendly kind of stuff. I don't know if GPU slicing will be in the initial GA. I don't know if you have any visibility into that, but the way, the speed at which they layer on additional features, I think is going to be pretty critical to the success for AVD on-prem. I don't see any
Starting point is 00:43:37 other way. If they can't do it quickly, it may not catch on. Well, that's the beauty of this iterative approach with the whole HCI stuff in general. So just to kind of segue back to Azure Stack HCI, the operating system to begin with, you are having a release cadence that's every year now. And in reality, it's almost every six months only because you're getting those public previews fairly quickly out to the public to even test this stuff.
Starting point is 00:44:05 So, yeah, we got GPUP, that's GPU partitioning, in a recent release of the OS itself. Well, now that allows your layers of services built on top of this OS to leverage it. So AVDHCI now, it's not publicly available yet, but GPUP is going to be available to it. Same thing with AKS hybrid. So your Kubernetes workloads may need a GPU, not for actual like graphics processing, but maybe doing some very specific floating point math or number crunching or goodness, any kind of crypto nonsense that I'm not well versed in. And if you have that, if you have those exposed to your VMs, okay, well, now you're able to let your containers touch the GPU itself. So there's a lot of benefit there. I think machine
Starting point is 00:44:56 learning was the big one with Kubernetes and GPUs on-prem. That was the kind of like, hey, we have these images that need to be processed and kind of pick out things with them. You know, like, is this car, is this human, is this dog, that type of stuff and what you can move there. Yeah. You, you mentioned at the beginning and then just, just a second ago, the Azure stack, HCI OS getting more regular updates, a quicker cadence there. Do you see an expansion of what Microsoft is doing with managed SQL into other elements of this infrastructure? Do you see a
Starting point is 00:45:38 vision where Microsoft does the whole management of the software on Azure Stack HCI, for instance, or other components within the system? So there definitely are things that are happening. I can't go into too much detail about the new features that may or may not be coming. Yeah, this is where my MVP hat comes on. I have to be careful with what I say. I can say that there's a lot of exciting things that I know are on the roadmap. One of the more recent ones was the release of,
Starting point is 00:46:12 and I think it's technically public preview still. I don't think it's GA yet. Just to link back to the support experience, the ability for you to have a remote support experience with Microsoft. You basically install a small agent that's connected to Azure that lives on your Azure Stack HDI host. Let's say you have to open up a ticket.
Starting point is 00:46:34 You open up a ticket. The support representative says, hey, can you allow me to dial in phone home, whatever the term they're going to use is, and they can get onto your machine for a certain amount of time and collect logs. They can only do kind of like read operations, not like set operations in PowerShell.
Starting point is 00:46:54 So it's a lot of gets, not sets. And they can collect data without you having to be the person to be collecting it for them. And that's great because you can significantly speed up the troubleshooting process from maybe one to two days to one to two hours. There was a lot of work going into the network desired state aspect of this. So again, you have your hardware deployed in your data center.
Starting point is 00:47:19 Day zero, you need to rack it. You need your network switches configured in a certain way. It's one of those things that you do once and you never think about again. And so you buy new hardware and you have to do it again. Or, oops, the server needs to be reinstalled. I need to do it again. That desired state is being worked on dramatically by Microsoft. The technologies that they're working on.
Starting point is 00:47:43 It's Network ATC and Network HUD. Those are the two very related products that Microsoft is going to be pushing in the next release of Azure Stack HCI. They will allow the OS to talk to the network switch, learn information
Starting point is 00:47:59 from maybe this switch port was misconfigured. Hey, and it'll send an alert to you and say, you need to go talk to your network admin and smack it around a bit to fix it. Just things like- Jeez, you guys are draconian over there. Hey, we can, that's a Dan Cuomo conversation
Starting point is 00:48:18 you can have at Microsoft, but I love the ability to be able to see all this. And then again, with it being all ARC connected, I love the ability to be able to see all this. And then again, with it being all ARC connected, a lot of this information ends up in the Azure portal, these alerts, these management tools where you can directly control and react to them in a single payment class. So, I mean, yeah, that's pretty neat.
Starting point is 00:48:41 And I think it comes back to then, I mean, you're a big advocate of Azure Stack HCI on-prem. I told you a little bit about our experience from a performance standpoint, and it's really pretty remarkable. The HCI product aside, you can go do a POC or get the license trials to go do that on your own test kit if you like. But it seems to be the Azure Arc thing is going to be broadly applicable to, as you said, a wide audience that's using Amazon or VMware or Azure or any of this stuff. You're not the sales guy, but is there a POC
Starting point is 00:49:23 or a pick up and go kind of way to start to experience ARC? The easiest way right now is to open up your browser and Google Azure ARC Jumpstart. Azure ARC Jumpstart is managed and maintained by a group of Microsoft employees that are brainiacs, super smart people. Basically, it is a single click and deploy into an Azure subscription where depending on the scenario you pick,
Starting point is 00:49:57 it will spin up your infrastructure for you in a lab environment. So we'll focus on SQL Server for a minute. You can pick the jumpstart for SQL Server managed instance through Arc. It will deploy a domain controller in the cloud for you as a small VM. It will deploy an AKS cluster in the same subscription in the cloud, and it will configure all the Azure Arc goodness extensions and what have you that you normally would have to use PowerShell or maybe the command line to enable.
Starting point is 00:50:28 It will run all these for you. It takes maybe like 30 minutes to 60 minutes to deploy the whole lab, but it will give you an environment that you can play around with like a sandbox and destroy it. And if you do, you can just delete it and recreate it with just another single click of the button. No, I love that.
Starting point is 00:50:43 I mean, NVIDIA has done a really good job with their launch pad uh set up for people that want to get touch some of the hardware like gpus and the bluefield smart nix and and other stuff because not everyone is going to especially when you start getting into the hardware going to be able to spin those things up and and can't go requisition a box of bluefield just to mess around with. So these labs, these companies have got to get smart to help their customers and prospects get hands-on and play. So that's pretty cool. I'll find a link to that and put that in the notes too so people can check that out. Ernie, this has been a great conversation. I love hearing from the practitioner standpoint, what you're doing, the things you're thinking
Starting point is 00:51:27 about. I think that's useful to the broader IT audience. You managed to get through this without being too much of a Microsoft fanboy. So that's good too. Yeah, try not to show. But no. You did a good job of playing in the and and delivering the information as you see fit and for for an organization your size to be as far into hyperv azure stack hci azure cloud all this other stuff is is uh pretty remarkable you guys are are out
Starting point is 00:51:58 there in the front so we we appreciate that and it's been a great conversation. Yep, great. Love doing this type of stuff. Anytime you want to chat, let me know. You got it, buddy. Thank you. All right. No problem.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.