PurePerformance - Successful Enterprise Monitoring Projects with Kayan Hales
Episode Date: September 21, 2020Successful Cloud Migrations, large scale Kubernetes & OpenShift deployments, making billions of data points actionable and enterprise-wide Citrix & SAP monitoring. These are some of the projects Kayan... Hales, Technical Manager at Dynatrace, and her colleagues at Dynatrace ONE help enterprise customers around the world to implement every day.We sat down with Kayan as we wanted to learn what really matters to many large organizations as they embark on automating monitoring into their hybrid multi-cloud environments. While we constantly talk about cloud native and microservices it was interesting to hear what the global team of Dynatrace experts is doing on a day-2-day basis. Kayan gives us insights how important it is to think about meta data, tagging strategies and automation before large scale rollouts and that one of the first question you need to ask is: who needs what type of data at which time through which channels.https://www.linkedin.com/in/kayanhales/https://www.dynatrace.com/services-support/dynatrace-one/
Transcript
Discussion (0)
It's time for Pure Performance!
Get your stopwatches ready, it's time for Pure Performance with Andy Grabner and Brian Wilson.
Hello everybody and welcome back to yet another episode.
It's always another episode of Pure Performance.
Welcome to another episode, Andy.
How are you doing on this another episode today?
It's a great another episode.
I wonder if I would start with saying this is not another episode,
but it's something completely different and you tuned in for the wrong podcast and now for something completely different how's everything going for
you andy it's very good and as i just told you in preparation today it's a special day for me
and i know it's going to be a special day coming up also for our guest in a couple of days
i got a special day coming up for me too special days all over all over yeah no other than that everything is good it is uh as of today after
recording early september um summer has accidentally disappeared over the weekend
here in austria at least it was warm until saturday and then a cold front came through
and it seems that's it with uh the summer fall is here kaput kap. It's kaput. Kaput. It's kaput. Yeah, exactly.
Yeah, we actually had a bit of a cold.
We're getting cold nights now
and then
here in Denver
and I looked at
the weather forecast
and we're going to have
a day in the 50s next week.
But that's Fahrenheit.
I don't know what that is
to Celsius
because, you know,
I'm a pig-headed American
who can't be bothered
to learn the metric system.
Don't you have your days
in your 50s every day?
I'm not that old.
Come on.
Come on.
Anyway, we have a guest.
A lovely guest, actually.
And I'm very happy to introduce her, even though I will keep it short because I'm pretty
sure she can do it much better than I.
Kayen Hales.
I'm actually, I guess i jinxed it earlier
because i'm definitely screwing up pronouncing her name um but i know that's what it is um it's it's
it's phenomenal having her on the podcast because i remember meeting her a couple of years ago when
she was a i think a guardian onsite with one of our
clients on the North.
I think it was Northeast.
And sat down, explained the Dynatrace product or helped hopefully, explained the Dynatrace
product and then helping the customer.
And it was very clear to see that while she was still, I think early early in the job. It was amazing how she kind of took the technical ideas
that the product realized and then pushing it over
and helping our customers implement it.
And now seeing her in a management role
within Dynatrace for the Dynatrace One Premium team,
which I think is really exciting.
But now I am stopping talking about her,
because I'm sure she can do this even better. Hi, how are you? And maybe you want to quickly introduce yourself
to the audience. Thanks, Andy. And I'll just say you almost got my name right. Well, you tried.
So I applaud the effort. But for the audience here, my name is Kayan Hills, as Andy mentioned.
I did work as a guardian at one of our accounts there in Northeast, as you mentioned.
And that actually was my first account and probably the toughest account and the best account.
So, yes, I am in management currently. And I do help the technical folks who report to me with their customers doing similar jobs where we're helping customers out, trying to get them to use the best monitoring practices.
And it's also cool.
I'm just looking at your LinkedIn profile.
You started through the PDP program at Dynatrace back in 2014.
I did PDP 319, which was the best class,
no matter what anyone else says. Yes, I did start through the PDP program.
What's PDP? You got to remember, this is all inside jargon. I don't even know what that means.
Right, right. Yeah. So PDP is the Professional Development Program, where they bring you in
young and fresh and wide-eyed
college grads, bring you in, train you in the technology that we use at Dynatrace and also in
the technology that we support. And then they let you loose and you're up to the customers and to
enable them to use the product successfully. I didn't even know we had that. That's awesome.
And I think the cool thing, that's also the reason why, you know, one of the reasons why we wanted to bring you on the call is because you have seen firsthand on site with, as you mentioned, one of the toughest accounts we had back then.
And kind of really seeing what problems people actually want to solve in the performance space. So today's talk is not about Dynatrace.
It's not about the different teams that we have to help our customers.
Well, obviously, we want to make sure that everybody understands
that we have a PDP program to educate our folks,
and we have a Guardian program where we send people on site.
And now you're part of Dynatrace One. So we can talk a little bit about it, but it's really interesting to see and hear from people like you, what are the real challenges
that large organizations really have when they are reaching out to monitoring vendors like Dynatrace
to say, hey, we need help. Because in the last couple of months, Brian and I specifically focused on CloudNavid.
Everybody's talking about Kubernetes.
It's the cool new thing.
And I'm pretty sure we see this a lot.
And this is where monitoring helps.
But I know there's much more out there.
And this is the reason why we really wanted to get you on the call and get a little overview
of what else is happening out there.
What are people trying to
solve and what challenges do they have as it comes to performance engineering monitoring
alerting and all these things that are relevant for us so um yeah and the question is go ahead
yeah the question is are you up for the challenge can we can we ask you a lot of questions yeah
absolutely i'm up for the challenge okay so
the first question that i have for you or the first kind of area and i know we talked about
this in preparation is that uh it seems now we talk a lot about the cloud yet a lot of people
are not in the cloud yet uh i think that we see a lot of organizations
that are trying to move into the cloud.
Cloud migration is a big topic.
Can you fill us a little bit in on what you see out there,
especially as people are moving into this new world
and what you see there where people are getting challenged
or what problems they want to solve?
Right, definitely.
Yes, I've definitely seen customers make that effort. are getting challenged or what problems they want to solve. Right, definitely.
Yes, I've definitely seen customers make that effort,
some customers quicker than others.
We've seen some customers where they're more interested in the next shiny object.
It's like, oh, yes, we have AWS.
Let's just go straight there.
Or, okay, now Azure has this new feature.
Let's just go there.
So they're kind of all over the place in a sense.
Then we have other customers or
this other camp of customers where everything's very rigid.
Where it's like, no,
we can't change our app,
no, we don't want to make any changes.
Yes, we know the Cloud is this great big thing,
but we don't want to do it because we don't want to mess up our apps.
What I found is the customers that are more rigid,
it's because they don't understand their application.
So they do look to us to help them figure out what exactly is their app doing.
And then the ones that are more attracted to that shiny object,
it's just exactly as it is.
It's like this new feature is promising to help them to do this thing.
And so they just want to try it out.
But we do have a customer
that we've worked with recently
that did a mass migration to AWS.
And this customer
is such a great example
because they actually had
a whole strategy
behind the migration.
So they had Dynatrace installed before that,
but they had their own strategy in-house, right?
A strategy that they worked out internally.
So they talked about cloud governance,
how can they optimize the network?
How can they leverage the support
that AWS is providing for them?
And make sure that they work directly with AWS in order to do that migration.
So that when they bring in monitoring tools, we just serve to kind of bridge the gap between what's in the legacy environment and what's in the cloud.
For example, doing that comparative analysis.
So okay, your response time is,
let's say 10 milliseconds in your legacy environment.
We want it to be that or better in the cloud environment.
So you want to pull out your monolithic app
into those different microservices and have a strategy for that,
bring it into the cloud and make a strategy for that, bring it into
the cloud and make sure that the requests from the legacy into the cloud
environment are lining up and that the response time lines up that way. So I
think customers that actually have a well thought out plan are definitely
more successful and they use monitoring tools to their advantage
to do that comparative analysis between performance on legacy platforms and performance in the cloud
hey and this is you brought up this example with you know monolithic applications
on-premise moving to a microservice architecture in the cloud have you seen this always the case that people are
actually really kind of replatforming and re-architecturing the app and breaking them up
or do you also see examples where they just still do the traditional lift and shift and i would i
would assume that you know the requirement would still be the same you want to make sure that the
application runs you know as least as fast if not faster in on somebody else's hardware but do you see maybe from this customer or others
more really the full we are re-architecturing for the cloud or do you also see these examples where
they're just lifting and shifting yep i've definitely seen both. Going the microservices way is recommended.
I think that's the best way,
but we've definitely seen customers that they
just want to do a one-for-one migration.
They want to take the entire monolithic app
and put it in the Cloud.
That tends to take a long time.
I do remember one of my very first accounts,
that customer tried
to do that lift and shift that you were talking about. And two years into the account, it
still did not happen. Just because to move an entire app that is so big all at once and
having so many teams involved, so many people that are interested in different performance metrics and so on.
And the timing has to work out right.
The change records have to be in correctly.
The firewall rules have to be in place and all of that.
It tends to delay things.
So definitely we've seen both,
but I do think going the microservices route is recommended.
Yeah, Andy, I think we saw this
or we had someone on from
AWS services on a while ago.
I think that was the conversation
where
their whole thing was they won't even
lift and shift.
Their whole thing was like,
our services
are here to help you
move to the cloud but
convert over to anything
else but lifting and shifting your monolith
does that ring a bell andy i kind of remember that conversation it does yeah same point came
out like uh kane you said you know it's two years and it's still not quite working and in two years
you could have had some or all of that converted over to something possibly more efficient,
only leaving the monoliths for things that made sense,
but everything else you could have done maybe more efficiently,
maybe more cost-effective.
Two years is a pretty good length of time,
and to still be struggling with a shift after two years,
I can see why people recommend against it, because it just seems like you just lost two years, you know. Right. And the thing is, you can still have a
functional app being hybrid. You can still have portions of the app in legacy, portions in the
cloud, and still be hybrid. You just have to break up the application into the logical functions,
and that creates the microservices, and put those in the cloud or make that gradual change
from legacy into the cloud using microservices.
And a lot of customers will actually see better performance in the cloud
just because those apps are broken up and you're able to function
or I should say focus on that specific part of the app
that may not be working.
And to say the least, if you're moving bits and pieces,
you're taking small chunks at a time with lower risk.
You're learning parts of the process as you're doing that.
So as you get to more of it, you have so much more experience.
It almost feels like the parts and limit is
that was that the company from the phoenix project project that was going on and on forever as they
were trying to get it all right you know it's like a little bit here at a time and those will
work and then you'll come back you know later on and make some better tweaks to them that you
learned on the the last parts but yeah anyhow sorry it just really struck me when you were
talking about that it It's great.
Thanks.
Yeah, absolutely.
And I do think, you know, the benefit of doing a little bit at a time is, you know, you do get the opportunity to understand your app.
Because like I mentioned before, some of our customers, they don't quite understand what their app is doing.
So when you break it up into those functional bits and pieces, they actually get a good understanding of what their app is doing and say,
oh, this part of the app does this and this part of the app is, you know,
for user experience and so on.
So I do think there's that benefit there.
And I think, Brian,
I think you mentioned earlier a podcast we had in the past.
I remember that discussion and I believe
one of the arguments from against lift and shift was if it if you just lift and shift an application into the cloud then
you don't really reap the benefits and maybe then people realize well why did we even lift and shift
it's not faster than before well how should it be faster are you just running on a different hardware
stack and i think that was also the argument of
people want to have a good experience
with moving to the cloud,
but in order to reap the benefits of the cloud
and having a good experience
is that you are actually adopting
to these new architectures,
leveraging the elasticity of the cloud,
leveraging cloud services for things
that you may have built and operated in-house earlier,
like a database, right?
I think that was also one of the reasons why they definitely favor
re-architecture and re-platforming apps as they move to the cloud.
One question, though, you said you compare before and after.
Is this, were you working then closely with performance engineering teams,
load testing teams that were actually able to simulate and put load on these systems simultaneously? That means you still had the old system and you had the new system in place and you were comparing and then you gave the go when you were confident that the new system is faster?
Or was it done more like you had data from the old system, you moved over that the new system is faster or was it done more like you had data from the old system,
you moved over to the new system
and then started comparing on production traffic
and then kind of making final tweaks
to the new cloud-based system?
Yep, so I'll say that what I've seen one customer do,
which is separate from the one
that we were talking about earlier, but what I've seen one customer do, which is separate from the one that we were talking about earlier, but what I've seen one customer the application and do another load test on the same
day at the same time so they can get comparative data for the performance so i i've seen them do
that in the lower environment first so they may start off with the development environment
and if that looks good if everything looks okay then they move on to, say, the test environment or the load testing environment, and then they'll move on to production.
So that's the method that I've seen so far for customers migrating.
And that has worked out so far.
The performance metrics that they track would be obviously response time, but also failure
rate.
CPU consumption is another big one,
and any errors or exceptions that's coming in
would be those that they would typically track.
Now, moving to the cloud and being in the cloud,
do you see additional data that people look at
from a performance engineering or monitoring and learning perspective
like are there are there new uh i don't know new metrics new something new or different to
the on-premise systems that people always ask for like hey i need to have visibility
on this because otherwise i fly blind or so. I've actually seen it the other way,
where on-prem, they're looking at so many metrics,
trying to build so many reports for upper management
because, at least based on what I've seen,
they don't quite understand the app.
So they're trying to just produce many, many different metrics
and try and figure out,
okay, since this metric is down, that means something is wrong or doing it in that approach.
Whereas when they move to the cloud, they actually focus on just the metrics that are important
and be able to use that to make decisions that affect the application. I've actually seen it
the other way where they actually end up using less metrics when they're in the application. I've actually seen it the other way, where they actually end up using less metrics
when they're in the cloud.
Yeah.
Andy, I was just thinking with that whole,
or in K2, sorry, but this whole idea of moving to the cloud
and changing the metrics kind of falls into the SRE pattern,
where if you think about at least the Google defined SLIs and SLOs,
the idea is you're not looking at things like CPU.
You're not looking at things like exceptions.
You're looking at uptime, response time for the end user.
You're looking at what the final product is looking like to determine then if you need to take action
and look at stuff in the back end.
And I think the cloud, moving to cloud,
moving to these scalable architectures
allows you to do that easier
because if you're on-prem and your own data center,
you have limited resources.
You have to maybe rack a new server,
try to fine-tune your VMs to get something out if an emergency happens. Whereas while we know we don't recommend throwing hardware
at the issue in the cloud even, at least you have that scalability and it's not as much of a scarce
resource where you're looking for every little metric and every little number to indicate,
oh my gosh, something might fall over and what are we going to do?
Because you have a lot more options.
Obviously, the best option is then to go ahead and fix what's causing the problem.
But you have, I'd say, the luxury maybe of focusing more on the end user and less on
a CPU number, which makes things, in my opinion, at least gives you more of a customer-based
focus.
Yeah.
I would have thought from it, I would have explained it from a different perspective.
Because, Kayan, you said that typically people on the on-premise side don't really know what
these apps are doing.
So probably they've been around for a long, long time maybe performance wasn't that uh important back then right and nobody really
thought about what's actually important so in order to get some type of visibility then
well let's get everything we can every single metric and then we figure out which metric might
tell us something but still we we collect everything because we don't know better and
it was never a requirement to actually define what's important and i believe now and brian this is goes definitely to your
point now when you are architecting and developing cloud native applications then one of the
requirements is to figure out what are my slis and my slos that are important to me because they're
important to the business?
And maybe what handful of other metrics do I need from a technical side
to get early warning indicators that something is wrong?
And so instead of doing the shotgun approach
of getting everything of an unknown system
because I need to figure out if case something is wrong,
then I want to have all the data.
I think we're moving more towards a,
I know exactly what's important to me
because we thought about this from the start
and therefore these things may get easier even
because it's more well-defined.
Yeah, exactly.
And I definitely agree that scalability is very important
and also capacity because once you
move into the cloud, you can put in some automation where if the throughput or if the load is
at a certain level, then you can add in another node automatically.
So those types of things are what customers tend to look at as well.
Very cool.
So we talked, can we go back to the, talked about a lot of data.
I know you have in preparation, you know, we looked at some of the things that you've
been doing and the kind of the belief of we need all the data in the world
is obviously something that you know people may feel more more safe if they have a lot of data
but i believe just collecting data for the sake of collecting data so we have data in case we
need data but we actually don't even know what the data tells us and we and and who even needs
that data is obviously a problem that, you know,
it's not actionable, let's call it that way.
And I believe, I know you have experienced this with some of the organizations
we work with.
It's just too much data, but it's not really actionable.
Is there kind of a trend that you see,
or is there maybe certain organizations
that are better in actually defining
and asking for actionable data versus others?
And also if you approach people
and you find out there's just too much data,
how do you make it actionable?
So I definitely have worked with customers
that they just want all of the data for
sure.
But as we look at the different requests that's coming in different transactions that's coming
in, there really isn't a need to look at each and every single one.
Right?
You what what what you want to do is identify any patterns or any trends over time.
So you do that with visual reporting.
So you can use dashboards
or any other type of visual reporting
to present that data to the forefront to say,
okay, my application was performing
consistently poor over a week.
Or you could even look at it from a comparative analysis to say, okay,
last week it was an acceptable response time, but this week is not.
So it's not necessary to grab all of the data, but it's more important to analyze the
data over a period of time to identify patterns.
And when you work with these teams,
do they typically know?
Like you mentioned the term constantly slow
or acceptable performance.
Does everyone have a clear definition
of what metric they actually would look at to define performance
and what their acceptable threshold is and the criteria or is this something that is not always
given that people actually know what's what I would say most customers tend to be reactive where they're only looking at the data, even though they they may try to use that as a baseline, but it's not the best baseline, right? It's not the best analytics that they can do. So, what we try to encourage them to do is, you know, look at the trends, look at the patterns, and have the most traffic. Let's say, for example, if it's an e-commerce site, right, it would be Black Friday, for example.
So you would look at that time where you have the most traffic and use that to measure your performance.
So if your application can handle that much load, then it's time to scale up or time to
consider capacity, time to look at different ways to optimize your code in order to handle
more.
When you work with these accounts and they ask you we need this type of dashboard we need this
type of data i mean i know you're advising a lot because you see you're doing this for many
accounts and i know you have we have a lot of different diamond trace one teams that are doing
this and i'm sure you're sharing best practices but what i would be interested in we over the
last couple of years we've talked a lot about data democratization, meaning opening up data to more people, making it easier to access, meaning giving developers data that they need, giving business, giving performance engineers, giving everyone access the the data we are capturing is actually put on dashboards and these dashboards
are then really used and circulated or made available to a large number of people and also
different types of groups or is it still more that you are just we're just building dashboards for
i don't know the platform team the performance But other than that, nobody gets access to that data.
Oh, yeah, we definitely have cases where we build dashboards as visible to everyone that's logging into Dynatrace
just to make sure that the performance of the application is visible for everyone.
We definitely have some customers that would rather keep their dashboards
specific to their team or specific to their group.
But for the most part,
we have just general dashboards
that can provide insight for everyone.
So we also do funnel dashboarding
where you have just one general dashboard,
but then you can click in on the different components
to open up like a sub dashboard or interconnected dashboards.
And that has proved to be very, very useful because you can see everything that you need in just one space.
You only have one link that you need to access with all of the metrics that may be important or relevant.
And we, of course, advise on that too.
And then based on your specific role or what you're interested in the company,
you can drill into that specific dashboard
and look at the metrics there.
So some may be role-based.
So we do have charts and metrics
that are specific to developers
versus application owners versus database guys
which would be very useful for them so yes we definitely do that
and and when you said you have to finally you said you started with saying we're building
funnel dashboards now when i hear funnel i immediately think of klaus ensenhofer because
he always talks about e-commerce conversion funnels
in your case I assume you talk about a funnel more like as a journey from I have an high level
overview and then I can either go into component ABC or maybe I'm a developer and therefore I'm
interested in this is this what you mean with funnel or maybe maybe I just didn't get this
right what do you mean with funnel yep that maybe I just didn't get this right. What do you mean with funnel?
Yeah, that is exactly that.
So we have a overview, high-level dashboard. Maybe it's focused on a group of applications that's tied to a specific business unit.
So they would go in and they would see the different, maybe one chart per application. And then for a specific app, they can drill in
and maybe look at infrastructure metrics
or application metrics or performance metrics overall.
And then based on what you're interested,
you can drill in to infrastructure metrics.
Let's say you're specifically an OS guy,
you would drill into infrastructure metrics.
If you're on a load test team,
you wanna look at performance metrics.
If you are focused on user experience, you'd wanna look at the application metrics. If you're on a low test team, you want to look at performance metrics. If you are focused on user experience, you'd want to look at the application metrics.
And even going further, like for example, if we pick application metrics, you can go even further
into that to look at how different geographies are performing for your application? What requests are users clicking on the most? How do you know
if a user actually reached the thank you page on your e-commerce site? Those are the things that
would be more of interest to those folks. Very cool. So that means we're advising or
you're advising on having an overview that basically aggregates data on an application level
so that means if you have a little tile for an application it would be something like
availability metric overall performance basically you know is it is it is it up or not i would
assume right or something like that or does it make revenue and things cool. And then drill down from there. Do we share these practices outside of the
Dynatrace one kind of group and advising customers?
Is there any I don't know.
I love my blogs. I love my YouTube videos.
Is there anything where we share some best practices more publicly?
Or is this more just
for Dynatrace One, for customers that are reaching out to Dynatrace One?
I would say from our perspective more so for customers reaching out to Dynatrace One.
So whether they do live chat or Dynatrace One Premium, they would get the same best
practices.
Perfect.
And of course, I know you have your blogs and so on.
Yeah. No, but it's great that you answered it that way, because that's a call to every
Dynatrace customer that is out there. Please either open up your chat in Dynatrace and then
reach out to the Dynatrace One team and say, I have just heard this blog post and I heard this podcast and we should talk about some dashboards, best practices.
That's good to hear.
Perfect.
Yep, absolutely.
Coming back to the cloud migration and the topic of re-architecturing
for the cloud as we did lift and shift.
I want to touch upon a topic that, as I said earlier, Brian and I have been
doing a lot of sessions or episodes on recently, the whole Kubernetes,
OpenShift, Cloud Foundry, the whole new container-based platforms.
I do assume that we do see this a lot in our accounts, do we?
Absolutely, yes.
Definitely, we initially saw a lot of Docker,
but lately we've been seeing a lot more Kubernetes
in our environments with the customers that we work with.
And do you see, is there a, I don't know, a trend towards Kubernetes on-premise or running it in the cloud as a managed service?
Is there a trend that you see of Kubernetes versus OpenShift or Cloud Foundry?
Is there kind of depending also on the type of organizations we work with any any kind of clear indicator or
let's say if a customer knocks on the door and you know they are in health care would you immediately
assume or guess based on data we have from others that yeah they are using probably open shift on
premise is there something like this we see out there I would say for the customers that we've worked with so far,
initially there was majority Docker and PCF,
Pivotal Cloud Foundry.
And then they would deploy using the Bash,
using the Bash deployment or also Pivotal Web Services.
But lately we have seen a shift from that into Kubernetes.
Kubernetes either in any of the cloud platforms,
more so than on-premise with our customers.
And they also offer monitoring to customers
who use Kubernetes, either through runtime or build time
so that the application owners can control
how they want their app to be monitored.
So it's definitely more hands-off
where we let the application team choose
what they want to monitor, but at the same time, we still help to provide the views that they need to see through dashboards
and most definitely help them with alerting, making sure it's being routed to the right person,
especially if they're using ITSM systems.
So that means, are you then advising the application teams to, I don't know if they want to have, I don't know, pod-based monitoring or code-level monitoring,
and they want to get the right dashboards, that they need to put some metadata on the containers,
whether it's tags or environment variables that is then triggering or allowing the right dashboards to show the data?
Is this also what you advise, like how they should build the containers or how they should
tag it?
Is this stuff that you do?
Yes, absolutely.
So we do suggest certain naming conventions that they should use.
Definitely need to add some metadata so that we can pull in the right information to populate
the tags.
And then the tags really drive everything.
So the tags drive the management zones, which drive the alerting and the dashboarding.
So, yes, we definitely allow them to or suggest to them what kind of metadata they should put on their pods.
And I remember now a discussion we had last week.
Or was it this week?
I don't remember.
About, you know, we have Perform coming up in February and
we're doing hot days, hands-on training days.
And I believe you suggested, you know, let's do a hot day session on, I think you called
it pre and post one agent deployment considerations, best practices around tagging and metadata.
Because I think this is a really important and we all just learn about how important tagging is
and what the best practices are in these very large dynamic microservice environments, correct?
Oh, yes, absolutely.
The first thing I tell a customer as soon as they deploy the one agent, let's talk about tagging.
That should be the first thing.
Aside from adding in host groups, definitely.
Because tagging is really what's going to drive everything else.
And it's really important to make sure that those are auto tags
because no one wants to spend hours creating manual tags.
We want them to be auto tags, and we also want them to be relevant.
We want to make sure tags are going to drive who needs to get alerted,
who needs to see the views, what kind of views we need to create.
We also suggest creating specific tags for a blue-green deployment as well, so that there
is no confusion about what's active and what's passive.
So tagging is crucial and there's best practices that we suggest for tagging your infrastructure
versus tagging your application services versus tagging your application in terms of real user monitoring.
Because all of that does play a part in the alerting.
The alerting piece is incredibly important, so it's very important to tag your application appropriately.
Kane, it almost sounds like what you're saying is that monitoring shouldn't be an afterthought,
but something that should be
built in from the ground up, right? And I'm saying this kind of with a wry smile on my face,
because this is the kind of stuff that we all understand. And besides hearing from you,
we hear from a lot of people that this is something you don't just, oh, we're going to
buy a tool and toss it on. Even if you're prometheus no matter what you're using right you have to plan these things properly to really get the benefit
out of it i mean you're talking about tagging right now right all the cloud providers have tags
you know aws tags the azure tags the gcp tags tags are everywhere setting up for monitoring
so that you could leverage these things is so critical.
And I'm really glad you're bringing up this point because it's just driving home the concept that,
you know, don't just, the old fashioned style, this is going back years to when I was first starting, right, of get a monitoring tool, turn it on, you have five servers, you know,
you can handle everything. Nowadays, you can't, there's no way to know what's what. So this is,
this has to be baked in early, has to become part of the best practices.
It has to become part of the thought process before you even start moving.
Do you find on engagements as you're planning those shifts into the cloud, whether it's a lift and shift or migrating to Kubernetes or whatever it might be are you all participating or finding people asking advice
at the planning phases on how we're going to bake good monitoring into this at that point
or does that still coming in a little bit later but the successful ones are at least doing it
before they go live what's your experience with that the successful ones definitely are the ones doing it before they go live because we have had some customers where they go live with their application.
They want to add monitoring afterwards.
And then based on the results of the monitoring, they find that they need to re-architect their application all over again.
So you do save a lot of time when you bake in monitoring from the very beginning.
When you're setting up your app, you're thinking about,
okay, how is this application name even going to show up in the tool?
How should I name it appropriately?
And how should I label my entities appropriately so that it shows up the way you want it in the tool
so that you don't end up doing a lot of configuration after the fact?
So naming conventions are really important. And just going back to tagging for a second,
it's a case where everything needs to be tagged.
So everything needs to have a label.
Otherwise, it gets lost in the wind, right?
Can you imagine when you have five blank pieces of paper?
Let's imagine that the blank pieces of paper are servers
and you have no idea what each server does
because you didn't label it.
And, you know, humans, we forget.
So it's very important that we label things appropriately
so that we can come back and say,
ah, yes, this is what this does.
So this is who it needs to go to.
This is the person who's interested in this
because of the label.
And I would further suggest too that
even if you're not moving to one of these complicated
systems you know let's say you still have more of a legacy or for whatever reasons a three tier or
you know that five server setup i was mentioning it's still a good idea to do these labeling as
just to start practicing for when and if you do these migrations you'll have that sort of ingrained
and it's you know if people can
kind of keep track of a small system but getting those practices and making it a habit early
is a good thing i think absolutely yes i agree with that so big lesson learned for everyone
it's about organizing the data right in the end you can you can collect data collection is easy
you know there's so many tools out there that you can use, but it's about thinking
how do you organize the data and the organization happens to metadata through
tags and is there, um, Brian, you mentioned AWS tags and so on.
Are there a best practices from a meta data and taking perspective of these
platforms like AWS, Kubernetes, OpenShift,
and so on, that have been kind of established as a standard?
Do you see this, that maybe we can all kind of learn about so that we don't have to reinvent
the wheel all the time?
Or maybe not that every vendor has to come up with their
best practices but maybe there's something more you know maybe something has already been defined
somewhere i don't know i think in kubernetes there's at least some annotations that are
kind of not defective but quasi standard but is there anything more that you see
any type of standards that are evolving around this? So definitely, I would say bringing in the tags that's already available in those platforms.
With Dynatrace, we can definitely do that, as well as making sure that you set your metadata
or your properties, environment properties, on the application itself, because all of
that will be automatically imported, and then we can make use of the data that's already there,
as opposed to having to go back and then reconfigure in the UI or in the Dynatrace UI
or in the monitoring UI. It's best to just make sure that everything is configured on the app,
so that if there are any changes that will automatically
be imported so that goes for metadata for environment variables properties process
properties i should say as well as any tagging that's done on the cloud platform itself
very cool um just a reminder for everyone that is listening, right?
I mean, these are obviously a lot of best practices that we have.
And I want to highlight again for folks that are interested in learning more about this
and also want to do it hands-on, the Perform 2021 is coming up virtual,
but we also have hands-on training days.
And that's going to be a hot day session
dedicated to to all of these tagging management zones and all that stuff so definitely shout out
to that session in case you're interested in i i would like to talk about one more thing or actually
kind of figure out if there is one more thing to talk about but i believe there is now we talked
about cloud migration we talked about kubernetes and the new cloud native world in the platform world we talked about a lot
of data making it actionable um but we talked a lot about kind of these new technologies
what else is out there i mean i'm pretty sure you see also other systems that
may not come to mind these days if you google observability which is the
kind of the hype word these days are there a lot of other systems that you see large enterprises
especially asking for to get visibility in from a monitoring perspective
i would say um citrix is a big one that enterprises are asking for visibility into right now.
But we do have monitoring for Citrix where we can do that through what we call an extension.
But the benefit of monitoring Citrix is that we can provide a lot of metrics that will help to understand what exactly is going on with Citrix,
what exactly you should be paying attention to.
So for example, you could look at database transactions,
we can look at active sessions,
we can look at app sessions, latency, and so on,
specifically for Citrix.
So that's definitely been a big ask from enterprises
and then we definitely have that capability.
So Citrix, obviously, big technology for large enterprises.
Is this mainly around, again, I know Citrix,
but mainly through Nestor,
but we typically talk about other things with Nestor
on the podcast and on webinars and on blogs.
But in this case, Citrix
is a lot about desktop virtualization, I would assume, right? So it's really about enabling the
employees of an organization to really work efficiently with the virtualized applications
and virtualized desktops, correct? That is correct. Yep. so we could do citrix virtual apps or citrix and app citrix and desktop
all those options very cool andy i hope we don't have to give nester another pair of shoes because
you brought them up it just comes to mind whenever the word citrix comes up obviously right
uh it's just the way it is yeah uh if you're listening we love you
i'm pretty sure he gets his next pair of sneakers anyway because i'm pretty sure once we roll back
to uh being on stage again at some point after covid uh he will be on stage again and even i
think speakers that do virtual engagements i'm pretty sure we'll get sneakers too sneakers for speakers that's the program so uh to kind of round this up so it's interesting for me as a confirmation right it's
not only the hot cool new thing that uh is important but it's also these technologies
that we may sometimes forget because we always talk about Kubernetes. Citrix is one of those.
I know we, do we see a lot of SAP out there?
Because I know we have SAP support too.
Yes, definitely we see SAP.
Another big one that came to mind is F5
and IBM, IBM MQ is another big one.
So definitely those technologies
where it's probably not as popular
as your typical Java or.NET.
We definitely see those where big customers are asking for support
for these technologies.
Very cool.
I think, I mean, I'm pretty sure we could probably talk a lot about best practices and
other things now technologies but overall i mean for me this was extremely insightful and especially
your your scenarios you know we talked about earlier with the cloud migration i think i found
this really fascinating because we still i think are just in the beginning of people migrating to
the cloud and knowing that uh those that actually have good planning and not just to lift and shift but to
really re-platforming other ones that are that tend to be more successful it's great to hear
and it's also great to hear that we have a strong dynatrace one team that supports and consults our
customers in how to use dynatrace for pre, post,
and also I think during their migration.
That's awesome.
Is there anything else we are missing?
Is there anything we want to get out to people that are listening?
They may or may not be Dynatrace customers.
So anything from a monitoring side, is there anything where you say,
well, I wish more people that look into
monitoring that look into performance engineering they would know about this so it will be easier
for us i don't know something comes to mind that's a good question let me think about that for a
second because there might be maybe you always run into similar issues that i don't know maybe
it's not the right people that are actually interested in performance
or is it the big challenge to actually sit down
and figure out what is actually important,
what needs to be monitored.
Nobody really have given me enough thought
on what it really is that they want to get out of monitoring
besides big dashboards.
Right.
I would say I do wish that customers would take the time
to understand the application.
And I know at customer sites, for some customers,
you have a large turnover.
So people are going in and out.
So the folks that know are leaving and new folks coming in,
they probably didn't get the correct training or so on.
But I do think that customers should take the time
to really understand the apps
and also take the time to really consider the alerting strategy.
I do think that's very important.
And just like we talked about tags earlier, that's driven by tags as well.
But to think about when something happens in Dynatrace
or when Dynatrace picks up a problem,
who exactly should it go to?
It has to go to the person
that can actually
take action on the problem.
So definitely
alerting strategy
and understanding
the application
are two big things
that I want customers
to take away.
Very cool.
Yeah, that's great advice.
Brian, is there anything from your end?
No, I got mine in during the show.
You know, something pops in my head,
I interrupt everyone and jump in.
But I think there's a lot of really cool
and interesting things that were talked about today.
And thank you, Kay, for bringing them all up and
i really just um you know in in the beginning you had i got to drop that because i lost the thought
um so i'll go with another one i really like um you know how it was very apparent by what you
discussed the importance of preparing and planning for your monitoring,
no matter what kind of monitoring you're doing,
what tools you're using.
Of course, we'd love you to use Dynatrace,
but all of these tools, all the modern tooling requires the tags,
requires organization, requires planning.
Do it now.
How you're going to prepare or how you're going to monitor,
even as you're talking about how you're going to route it to the right team. So tags, all these things feed into that. So it's not something that should be thought of as an
afterthought. You know, we're seeing more and more, and Andy and I, you and I have seen this
since all the years we've been working in performance, the rise of the importance of
performance, from performance being an afterthought,
from performance testing being that thing that those people do
and they just hand us some numbers
and they don't know what they're doing anyway,
to it becoming more and more and more centrally focused.
And it's just really great to see that all happening
and really being taken seriously,
especially because of how critical of a role it plays
in these modern environments where, yeah, you have the scalability options,
you have infrastructure at your fingertips,
but as we know, if you just throw hardware at it again,
you have the same problem of costing more money,
and it's so much easier to spend money in the cloud without knowing it than on-prem.
So a huge piece of it, and Kay, thank you so much for sharing your experiences.
And I think it's also your experience of going through that program and now being in management
is pretty awesome.
That's a great accomplishment on your part.
So congratulations to you for just being awesome.
Thank you.
And it was really a pleasure to be here.
Cool.
Andy,
do you have any,
are we summer,
summer,
are we summoning the summer today?
Sure we can,
but we don't have to,
but I'll keep it short because I think I already reiterated to most of the
things,
but Kay and typically at the end,
I kind of summarize what I learned.
And I think it comes down to what just just said right it's it's very important that people understand that monitoring
is critical but more important is that they understand what to monitor how to monitor how
to organize the data also who is the beneficiary of the monitoring data who should be alerted
um who should get what data at which time and what level of data do we actually need
i also really liked the fact that we are you know seeing these cloud migrations as i just
said and helping people into these new platforms like k or OpenShift. But that we, on the other hand, see that this is a diverse world we live in
from a technology perspective as well as with a lot of other things in life.
And that we still see a lot of customers that are monitoring their Citrix environments
and that are monitoring their SCP, their F5s, their their mqs and so on so it's it's really important
that you know we think of monitoring holistically across the whole organization because not today
not tomorrow and not in three years will you just have kuben microservices in kubernetes we will
have a very diverse set of technology for a very long time. And all these pieces are very important.
And so we need to figure out what are the important metrics
so we can make the right decision in case something happens.
And I think you did a great job in telling us that
we as an organization have a great dynamic with one team
that supports our customers.
We learn from them across the globe, across different verticals.
And we have best practices and great teams
that can apply these best practices
to all the other customers.
And thank you so much for doing this job.
And hopefully, yeah, we'll have you back again
with more stories in the near future.
Thank you very much, Andy.
Awesome.
All right. Well, if anybody
has any
topics or anything they'd like for us
to discuss, they can reach out to us at
pure underscore DT.
Kay, do you do
LinkedIn or anything
that you want to
get any followers you want to put out there?
I'm on LinkedIn at KNHales.
KNHales?
Yes.
H-A-L-E-S, not H-A-Y-E-S.
As Andy might spell it, right?
You got it right. H-A-L-E-S.
I'll just do my own shame
and self-plug. September 4th,
my band,
we have an EP coming out on streaming it's
a compilation of stuff that wasn't released between 2004 and 2008 the first of three
eps of unreleased material so i'm excited about that and uh andy anything that you wanted to bring
no i think that's it um well actually no i have to bring up one thing because we haven't brought it up today. It's called Captain.
Go to www.captain.sh.
It's a new thing we're working on
and it would be great to have more people on that project.
But that's it.
Yep, and we have Perform is going to be virtual this year.
So I think there's been some,
I don't know if there's been any releases yet.
I know there's some stuff internally going on with that so far, but keep a lookout for that
people. That'll, there'll be some news coming out if it hasn't already hit the streets yet on
what's going on there. And I guess everyone, you know, you might not be able to get your unicorn,
but please remember to vote.