We are using it in an operational mode, when we have trouble easily getting the root cause, getting the application back up and running.
Based on that, the product has worked very well for us. We are happy with it.
We are using it in an operational mode, when we have trouble easily getting the root cause, getting the application back up and running.
Based on that, the product has worked very well for us. We are happy with it.
It's really opened our eyes. We had point solutions where we could see different elements of the stack, and Dynatrace ties everything together. Before, we could never get that full-stack monitoring. It prevents that, "Oh, it's your problem. No, it's your problem," type of an issue, and it allows us to get to that problem.
It also helps us get us the context of the customer experience. What's the business impact of those problems? And we've never had that before. That has been good.
Where we are struggling is being able to pull that information out and combine it with other contextual information that we have in other sources. Mining that data in a big-data environment, and joining it together and coming up with larger types of analysis on it. Big-data types of issues. We're still blazing a trail, trying to figure that out. But it's not as easy as some of the other things we've been able to do with the product.
Very stable. Very happy with it.
We have a lot of our infrastructure on it, so it's meeting our needs, for our enterprise. We have thousands of agents that are out there in over a thousand applications, and it's meeting our needs with that.
I think it's good. They are very responsive and get back to us. They try to give us workarounds and follow up with us. So, we're happy with that.
We have an infrastructure group and I'm more on the business-unit side, but I was part of our PoC as we brought it in, and stood it up. Generally, it was very easy to get it set up and get going very quickly. It was pretty easy. We used some of the Dynatrace sales team and the engineers to help us get it set up, but in short order, we had it going.
AppDynamics and New Relic were the other two.
We were never able to get AppDynamics working in our PoC. We couldn't get it working on our web servers. New Relic didn't meet some of our shortlist criterion.
Regarding the nature of digital complexity, I think the role of AI is becoming more critical when it comes to IT's ability to scale in the cloud and manage performance problems. It's because of the complexity and the number of elements that are out there, and being able to completely understand what the problem is. There was a good quote from one of the last keynote presentations here at the Perform 2018 conference: "Let's not chase $500 issues. Using AI allows us to go for those bigger issues," and look for more value, rather than worrying about all the little things that happen. AI would give us the ability to handle that low-level work, very quickly - the auto remediation - get that back up and going. It would buy us time to do higher-level work.
We've used a lot tools at our company, including siloed monitoring tools. Some of the main things we're seeing with them are gaps in the ability to handle emerging technology; things like single-page applications, Angular applications, single sign-on applications, those types of things.
When looking at purchasing an APM solution, we wanted something that was a proven leader. We looked at industry review rankings. Did it support the technologies we develop our applications on? Can it give us that full-stack view into our architecture? Can it tell us what's going on with the customer experience? Those types of things.
If I had a friend looking to adopt an APM solution, I'd really have him take a look at Dynatrace. It's an industry leader. We've had a great experience with them. It meets our needs. They're future-looking. Even though we're not where they are in terms of the capabilities they have, we know we're going to need those capabilities in the future. Great product.
The main use of Dynatrace is development assistance through to live deployment, then proactive monitoring. Afterwards, monitoring of the infrastructure and live site.
Being able to track issues found in live back through environments is particularly useful.
The obvious benefits are proactive monitoring, but there were the unexpected results we gained from it that brought additional value out of the product:
This is not a simple product. You cannot fire and forget. Maybe not a specific function issue that needs resolving, but certainly an area for consideration upon adoption of the product.
Training is required for all of the people who will be using it, and this should not be overlooked. I would even recommend nominating an SME in each of the three areas covered: user behaviour analysis, development, and infrastructure/operations support.
We have had some issues, mainly in the JS embedded in the page.
None.
We have used them extensively. They are very eager to help.
Reviewed other products, but never switched.
Fairly complicated, but it is an in-depth product.
In-house.
As a user, I can testify that we found an issue within the first week of ownership that has been costing us more than the entire license cost.
Do not ignore the training!
Yes, AppDynamics, New Relic, and one other.
Being able to dig into code to figure out errors and where response time is slow.
We have setup proactive alerts to notify us when certain conditions are met, such as when we are out of memory or high threads.
I cannot think of any off hand. They are continuously making the product better.
I've been using Dynatrace Application Monitoring and Data Center RUM for three years.
(version 6.0)
I'm not aware of any issues.
No, it is very scalable.
10 out of 10 as they are great to work with.
Technical Support:9 out of 10, but it could be 10 out of 10. They are very prompt and technical. They will work with you to fix any issues.
No, we reviewed other vendors and thought this one was the best.
It was simple to implement.
Dynatrace came in, helped install it and then showed us how to use the product.
We serve multiple customers and everyone wants to use Dynatrace. It has paid for itself. We can now figure out issues so much quicker.
CA was one other options we considered amongst others.
We use Dynatrace for a number of internal applications that we track in addition to API calls associated with the API engine. We have a partnership with Dynatrace and I'm a project manager.
We monitor critical internal applications including some public-facing applications. Internal transactions are being tracked and we get immediate feedback from the solution's monitoring which makes a big difference to us.
The value of this solution is in terms of the functionality, and every aspect of the hardware and connection-oriented signals that we get. We use most of the features on a daily basis.
Network monitoring doesn't seem to be a key focus of the company and if that were improved this could be a one-stop solution that would monitor the application. It would be quite useful in the data center environment as well.
I've been using this solution for four years.
The solution is stable.
The solution is scalable, we have around 50 users.
I am unaware of licensing costs.
It's a wonderful product and I would definitely recommend it. I rate this solution eight out of 10.
Regarding the use case for Dynatrace, currently, my company is working on a UK-based project in the telecom industry. I'm working as one of the leads on the Dynatrace side. The project has three vendors supporting it as it's a massive project.
My role in the project is installing new agents, giving support for any BAU issues that arise, and creating dashboards. My company works with different vendors, such as IBM and TCS. It's the vendors that support RUM (real user monitoring) and licensing. On the other hand, my company handles agent deployments and gives that over to the application team. Whenever there's an issue, I assist in finding out the root cause and fulfilling requests such as creating new dashboards, customizing dashboards, etc.
One of the most valuable features of Dynatrace is that it offers good visibility. It's better than other APM tools. You're not required to use a different technology when you have Dynatrace because it will work whether you're hosting it on Windows or Linux.
For example, Dynatrace can do full-stack monitoring for a particular server if you use it on the Windows platform. Whatever application is running, you can monitor it thoroughly, even the services, processes, and environment.
I also find the minimal configuration and one-time deployment of Dynatrace valuable and that you can do both infrastructure monitoring and application monitoring on Dynatrace. You can also restrict the feature. For example, if you don't want to do infrastructure monitoring, you can limit that feature.
What needs improvement in Dynatrace is its dashboard. Creating dashboards in Dynatrace is good, but compared to Grafana, which is integrated with Broadcom DX APM, the resulting dashboard in Dynatrace isn't as clear. The Dynatrace dashboard needs to be more graphic.
Since I'm not using the latest version of Dynatrace, I cannot share what additional feature I'd like to see from the solution. I would need to use the latest version first to see if there's anything I'd like added to it.
I've been working with Dynatrace for the last three years, and I'm still using it.
Dynatrace is a very stable solution.
Dynatrace is a scalable solution.
My team reaches out to the Dynatrace technical support team whenever there's a connectivity issue, an issue with dashboard creation, or a log that needs clarification. I used to contact the vendor via email and create a request for the Dynatrace support team.
Support is good, but it needs a faster response, so I'd give it an eight out of ten.
Positive
I prefer Dynatrace over other APM solutions because it's the best in the market, except when compared to the visualizations in the Grafana dashboard. Dynatrace has a host-based deployment, so unlike other solutions that require you to create supporting files and packages based on the application and server, in Dynatrace, as long as the application runs on Windows or Linux, that's it. You don't need to gather files or do anything else. Other APM tools require you to select the application type, and then you need to download the right package. You need compatibility when you use other APM solutions, which is a big headache, unlike Dynatrace, so I like Dynatrace better.
Dynatrace is easy to deploy. It won't take more than two minutes to deploy one agent, but getting approval, summarizing the changes, and the change request concerning the service ticket getting approved is what's taking longer.
I deployed Dynatrace, so it was an in-house deployment.
A different team handles the licensing for Dynatrace.
I evaluated Grafana and Broadcom DX APM.
I'm working with Dynatrace. I'm using an older version of the solution.
Within the company, eight people use Dynatrace.
The solution is straightforward to maintain.
Even if I have three and a half years of experience with Dynatrace and a total of eleven years of experience with other APM solutions, my only advice to new users or anyone looking into implementing Dynatrace is that it's a good tool. However, I still need to dig deeper into the solution to give more advice.
My rating for Dynatrace is ten out of ten.
My primary use cases of this solution are to understand how users are interacting with and experiencing applications and to quickly identify and fix problems.
Dynatrace has reduced our total headcount in operations and the mean time to detect and resolve problems. As a result, those challenging offline times are much shorter, if not non-existent, because of this solution.
The most valuable features are session replay, which allows for full playback of a user's experience; the AI engine "Davis," which does problem identification; and automatic mapping, which gives a visual representation of how applications interact host-to-host or process-to-process.
An area for improvement would be security. In the next release, I'd like to see more network-centric capabilities - Dynatrace is good at the network level, but I have to leverage other network solutions and integrate with them, but a holistic approach including the network as a one-stop-shop would be great.
Dynatrace's stability is solid - it performs updates very often, so it's always the latest and greatest in a good way.
Dynatrace has phenomenal scalability capabilities.
The technical support is phenomenal - they have a call program called Dynatrace ONE, which is like a customer success program on steroids.
The initial setup was extremely straightforward and fast. The deployment function was also super fast, typically just a few hours at most, with the right tuning.
When used appropriately and applied to the applications that are meaningful for businesses, the ROI is extremely high.
There's a perception that Dynatrace's value could be questioned, but this is down to a lack of due diligence on the front end. When done right, this product always gives good ROI and total cost of ownership.
Dynatrace is really good at keeping some infrastructure details and really good at the application level. I would give this solution a score of ten out of ten.
Our primary use case is the consolidation of observability platforms.
Looking at Dynatrace's automation and AI capabilities, automation is generally a great place to start. In products where there has been no observability or a very limited amount, the automation can give a great deal of insight, telling people things that they didn't know that they needed to know.
Davis will do its best to provide root cause analysis, but you, as a human, are still responsible for joining as many of the dots together as possible in order to provide as big a picture as possible. As long as you accept that you still have to do some work, you'll get a good result.
I have not used Dynatrace for dynamic microservices within a Kubernetes environment in this company, but I have had an AWS microservice cluster in the past. Its ability to cope with ephemeral incidences, as Kubernetes usually are, was very good. The fact that we didn't have to manually scale out to match any autoscaling rules on the Kubernetes clusters was very useful. Its representation of them at the time wasn't the best. Other products, Datadog, for example, had a better representation in the actual portal of the SaaS platform. That was about three years ago, and Dynatrace has changed, but I haven't yet reused the Kubernetes monitoring to see if it has improved in that regard.
Given that Dynatrace is a single platform, as opposed to needing multiple tools, the ease of management is good because there is only one place to go in order to manage things. You deal with all of the management in one place.
The unified platform has allowed our teams to better collaborate. In particular, because of the platform consolidation, using Dynatrace has made the way we work generally more efficient. We don't have to hop between seven different monitoring tools. Instead, there's just one place to go. It's increased the level of observability throughout the business, where we now have development looking at their own metrics through APM, rather than waiting until there's a problem or an issue and then getting a bug report and then trying to recreate it.
It's increased visibility for the executive and the senior management, where they're getting to see dashboards about what's happening right now across the business or across their products, which didn't used to exist. There's the rate at which we can monitor new infrastructure, or applications, or custom devices. We had a rollout this week, which started two days ago, and by yesterday afternoon, I was able to provide dashboards giving feedback on the very infrastructure and applications that they had set the monitoring up on the day before.
As we've only been using Dynatrace in production for the past month in this company, the estimate as to the measurement of impact isn't ready yet. We need more time, more data, and more real use cases as opposed to the synthetic outages we've been creating. In my experience, Dynatrace is generally quite accurate for assessing the level of severity. Even in scenarios where you simply rely on the automation without any custom thresholds or anything like that, it does a good job of providing business awareness as to what is happening in your product.
Dynatrace has a single agent that we need to install for automated deployment and discovery. It uses up to four processes and we found it especially useful in dealing with things like old Linux distros. For example, Gentoo Linux couldn't handle TLS 1.2 for transport and thus, could not download the agent directly. We only had to move the one agent over SSH to the Gentoo server and install it, which was much easier than if we'd had to repeat that two or three times.
The automated discovery and analysis features have helped us to proactively troubleshoot products and pinpoint the underlying root cause. There was one particular product that benefited during the proof of concept period, where a product owner convened a war room and it took about nine hours of group time to try and reason out what might be the problem by looking at the codebase and other components. Then, when we did the same exercise for a different issue but with Dynatrace and the war room was convened, we had a likely root cause to work from in about 30 minutes.
In previous companies where the deployment has been more mature, it was definitely allowing DevOps to concentrate on shipping quality rather than where I am now, which is deploying Dynatrace. The biggest change in that organization was the use of APM and the insights it gave developers.
Current to the deployment of Dynatrace, we adopted a different methodology using Scrum and Agile for development. By following the Scrum pattern of meetings, we were able to observe the estimated time in the planning sessions for various tasks. It started to come down once the output of the APM had been considered. Ultimately, Dynatrace APM provided the insight that allowed the developers to complete the task faster.
The most valuable features for us right now are the auto-instrumentation, the automatic threshold creation, and the Davis AI-based root cause analysis, along with the dashboarding for executives and product owners.
These features are important because of the improved time it takes for deployment. There is a relatively small team deploying to a relatively large number of products, and therefore infrastructure types and technology stacks. If I had to manually instrument this, like how it is accomplished using Nagios or Zabbix, for example, it would take an extremely long time, perhaps years, to complete on my own. But with Dynatrace, I can install the agent, and as long as there is a properly formed connection between the agent and the SaaS platform, then I know that there is something to begin working with immediately and I can move on to the next and then review it so that the time to deployment is much shorter. It can be completed in months or less.
We employ real user monitoring, session replay, and synthetic monitoring functionalities. We have quite a few web applications and they generally have little to no observability beyond the infrastructure on which the applications run. The real user monitoring has been quite valuable in demonstrating to product owners and managers how the round-trips, or the key user actions, or expensive queries, for example, have been impacting the user experience.
By combining that with session replay and actually watching through a problematic session for a user, they get to experience the context as well as the raw data. For a developer, for example, it's helpful that you can tell them that a particular action is slow, or it has a low Apdex score, for example, but if you can show them what the customer is experiencing and they can see state changes in the page coupled with the slowness, then that gives a much richer diagnostic experience.
We use the synthetics in conjunction either with the real user monitoring or as standalone events for sites that either aren't public-facing, such as internal administration sites, or for APIs where we want to measure things in a timely manner. Rather than waiting for seasonal activity from a user as they go to work, go home, et cetera, we want it at a constant rate. Synthetics are very useful for that.
The benefit of Dynatrace's visualization capabilities has been more apparent for those that haven't used Dynatrace before or not for very long. When I show a product owner a dashboard highlighting the infrastructure health and any problems, or the general state of the infrastructure with Data Explorer graphs on it, that's normally a very exciting moment for them because they're getting to see things that they could only imagine before.
In terms of triaging, it has been useful for the sysadmins and the platform engineering team, as they normally had to rely on multiple tools up until now. We have had a consolidation of observability tools, originally starting with seven different monitoring platforms. It was very difficult for our sysadmins as they watched a data center running VMware with so many tools. Consolidating that into Dynatrace has been the biggest help, especially with Davis backing you up with RCAs.
The Smartscape topology has also been useful, although it is more for systems administrators than for product owners. Sysadmins have reveled in being able to see the interconnectedness of various infrastructures, even in the way that Dynatrace can discover things to which it isn't directly instrumented. When you have an agent on a server surrounded by other servers, but they do not have an agent installed, it will still allow a degree of discovery which can be represented in the Smartscape topology and help you plan where you need to move next or just highlight things that you hadn't even realized were connected.
The user interface for the management functions is not particularly intuitive for even the most common features. For example, you can't share dashboards en masse. You have to open each dashboard, go into settings, change the sharing options, go back to dashboards, et cetera. It's quite laborious. Whereas, Datadog does a better job in the same scenario of being a single platform of making these options accessible.
User and group management in the account settings for user permissions could be improved.
The way that Dynatrace deals with time zones across multiple geographies is quite a bone of contention because Dynatrace only displays the browser's local time. This is a problem because when I'm talking with people in Canada, which I do every day, they either have to run, on the fly, time recalculations in their heads to work out the time zone we're actually talking about as relevant to them, or I have to spin up a VM in order to open the browser with the time zone set to their local one in order to make it obvious to them without them having to do any mental arithmetic.
Personally, I have been using Dynatrace since November of 2018. At the company I am at, we have been using it for approximately four months. It was used as a PoC for the first three months, and it has been in production for the past month.
The SaaS product hasn't had any downtime while I've been at my current company. I've experienced downtime in the past, but it's minimal.
To this point, I've not had any problems with the scalability, aside from ensuring that you have provisioned enough units. However, that is another point that is related to pricing.
Essentially, its ability to scale and continue to work is fine. On the other hand, its ability to predict the required scalability in order to purchase the correct number of various units is much harder.
Talking about Dynatrace as a company, the people I've spoken to have always been responsive. The support is always available, partly because of our support package. As a whole, Dynatrace has always been a very responsive entity, whether I've been dealing with them in North America or in the UK.
We have used several other solutions including Grafana, Prometheus, Nagios, Zabbix, New Relic, AWS CloudWatch, Azure App Insights, and AppDynamics. We switched to Dynatrace in order to consolidate all of our observability platforms.
Aside from differences that I discuss in response to other questions, other differences would come from the product support rather than the product itself. Examples of this are Dynatrace University, the DT One support team, the post-sales goal-setting sessions, and training.
We're yet to have our main body of training, but we're currently scheduled to train on about 35 modules. Whereas, last year, when I rolled out Datadog, the training wasn't handled in the same way. It was far more on request for specific features. Whereas, this is an actual curriculum in order to familiarize end users with the product.
In my experience, the initial setup has been straightforward, but I've done it a few times. When I compare it to tools like Nagios, Zabbix, Grafana, and Prometheus, it is very straightforward. This is largely for two reasons.
First, they're not SaaS applications, whereas Dynatrace is, and second, the amount of backend configuration you have to do in preparation for those tools is much higher. That said, if we were to switch to Dynatrace Managed rather than Dynatrace SaaS, I imagine that the level of complexity for Dynatrace would rise significantly. As such, my answer is biased towards Dynatrace SaaS.
In my previous company, it allowed a very small team to manage what was a very fast-moving tech stack. In my current company, it is still very early.
The consolidation of tools due to implementing Dynatrace has saved us money, although it's tricky to measure the impact. The list price of Dynatrace was more than the previous list price spend on monitoring tools because the various platforms had been provided as open-source tools, were provided through hosting companies, or had been acquired as part of acquisitions of other companies.
The open-source applications that we used included Grafana, Prometheus, Nagios, and Zabbix. New Relic through Carbon60 in Canada, as an example, was provided through a hosting company. Also, we acquired a Canadian company or had been acquired as part of acquisitions of other companies, AppDynamics, in a Canadian company, for example, with us in the budget of the previous company rather than our own company.
The hope was that Dynatrace through consolidation would release the material cost of the administrative overheads of tools like Prometheus and Grafana and the cost of hosting infrastructure for solutions like Nagios, Zabbix, Prometheus, Grafana, et cetera. This means that it is more of an upstream cost-saving, where we would be saving human effort and hosting costs by consolidating into a SaaS platform, which is pretty much all-in-one.
Dynatrace's pricing for their consumption units is rather arcane compared to some of the other tools, thus making forward-looking calculations based on capacity planning quite hard. This is because you have to do your capacity planning, work out what that would mean in real terms, then translate that into Dynatrace terms and try to ensure you have enough Davis units, synthetics units, DEM units, and host units.
Catching those and making sure you've got them all right for anything up to a year in advance is quite hard. This means that its ability to scale and continue to work is fine but predicting the correct number of various units to purchase is much harder.
The premium support package is available for an additional charge.
At this point, we have not yet integrated Dynatrace with our CICD tool, which is Azure DevOps. However, in the future, our plan is to provide post-release measurements and automated rollbacks when necessary. Even further down the road, there's ServiceNow on the roadmap, which we're currently bringing in from an Australian acquisition in order to try and promote the ITSM side of the business.
There is nothing specific that has been implemented so far, although there have been general degrees of automation. When we get Agile, DevOps, and ServiceNow in place, the degree of automation will increase dramatically. For example, automated rollbacks in the case of deployment failure or change management automation through the current state of the target system are being included in the ServiceNow automation.
The automation that has been done to alleviate the effort spent on manual tasks is still very light because I'm the only person doing the work. I generally don't have time to do the ancillary tasks at the moment, such as creating automations. It's mostly a case of deploying instruments, observing, and moving on. When we come back to revisit it, then we'll look at the automations.
My advice for anybody who is looking into implementing Dynatrace is to make sure you talk constantly with your Dynatrace representatives during the PoC, or trial phase because there is invariably far more that Dynatrace can do than you realize. We only know what we know. I'm not suggesting that you let Dynatrace drive but instead, constantly provide the best practices. You will achieve faster returns afterward, whether that's labor savings, or recovery time, or costs from downtime. Basically, you want to make sure that you leverage the expertise of the company.
In summary, this is a very good product but they need to sort out their user interface issues and provide a more logical experience.
I would rate this solution a seven out of ten.
The one that we recently scripted was just to see if an application was up, it was a very simple script. We had an issue with a vended solution at the university in which the application would just go down and the vendor had said that, "Oh, it never goes down, we have 100% uptime." We didn't have a good way to monitor that before. We said, "Our students our reporting that they cannot log in." With Synthetics, we wrote a very simple script that went to the landing page and tried to type and hit "enter," and that's it.
We have two-factor and some other things that prevent us from going too far into it, and we haven't figured out a technical solution for it, but it's a very simple use case of just making sure that the application is up.
Dynatrace itself is performing well.
We now know when the application is down, as opposed to students opening tickets and letting us know. So it's more proactive.
That it's always running.
From what I've learned today, here at the Perform 2018 conference, there are two things that I really want to see.
Number one, the thing that is preventing MSU from moving forward with Dynatrace right now is that we can't tag our customer traffic with a customizable tag. All of our students have a unique identifier and in AppMon we tag that and we can search by it very easily and it's very useful. But in Dynatrace, you can't yet customize and find people like that, so that's really preventing us. I heard that it's being worked on but I'm not sure when it's coming out.
The second thing that I'd really love to see is - I'm very impressed by all the new features that I've learned about at the conference - and one of the features is "impacted users." I would like to see a rate of impacted users. For example, how long has the problem been going on: 100 users in five minutes. Does that mean that in 3 hours if we don't get this solved, we're impacting x number of people? Understanding the rate at which the problem is impacting people would be a cool feature.
It seems very stable.
With the AI, it seems very scalable.
We're going through a proof of concept right now, so we work very closely. They're knowledgeable and we get the right person whenever we call.
We didn't have a previous APM solution. We didn't know we needed this solution until we saw it. It was literally students calling in with problem tickets.
I'm just a supervisor, so I sat with our technician during setup. I didn't do any of the actual work, but it seems seamless. It installed in about two minutes. I really wanted to see it. I have to go to the assistant director and, eventually, the director and try to explain why I think we need the new technology.
They went through a whole eight month process. I wasn't there, I'm a new manager, but I understand they had over a dozen companies. They had spreadsheets of all the pros and cons. They went with Dynatrace because it just has more features.
Regarding AI and the ability of IT to scale into the cloud and manage performance problems, we don't have the new technology yet, we don't have the new AI, we have the old AppMon and Synthetics. But it seems like it's very important.
We have used siloed monitoring tools in the past. The university is very old and very segmented and different departments use different tools and we don't all talk to each other. We still have this problem today and we're trying to get more user adoption for Dynatrace. But it's difficult.
If we had just one solution that could provide real answers, and not just data, I think that we would spend less time working on things that probably don't matter, like mundane routine operations.
To me, the most important criteria when working with a vendor is responsiveness to issues.
My advice to colleagues would be, do your homework. But, I would be surprised if anybody would beat Dynatrace.