Try our new research platform with insights from 80,000+ expert users
reviewer2587689 - PeerSpot reviewer
Cloud/Devops Engineer at a computer software company with 1,001-5,000 employees
Real User
Top 5
Offers broad-level exposure that has increased our efficiency
Pros and Cons
  • "It helps us meet our service-level agreements. It is integrated into our CI/CD pipeline. It enhances our operational productivity."
  • "Control-M has improved our organization's functions by supporting high availability and integrating with CI/CD workflows."
  • "The UI can be challenging for new users due to its learning curve. Additionally, there are some errors during automation. More detailed logs would be helpful."
  • "The UI can be challenging for new users due to its learning curve."

What is our primary use case?

I am using Control-M for cloud infrastructure and automation-related tasks. As a cloud engineer, my work involves scheduling, deploying, managing, and monitoring processes for infrastructure and workflows. It is integrated with a CI/CD toolchain as part of our DevOps culture.

I am using the cloud version. I am using Helix Control-M.

How has it helped my organization?

Control-M has good integration capability. It integrates well with all the solutions. It also has good reporting capabilities.

Control-M has improved our organization's functions by supporting high availability and integrating with CI/CD workflows. It helps maintain high availability and manage workflows across the production environment, increasing our productivity.

Additionally, Control-M has offered broad-level exposure that has increased our efficiency. Our workflows run smoothly. Everything is easy. We have had very positive feedback.

Control-M is fine to integrate with our DevOps toolchain. It is neither difficult nor easy.

Control-M made it more simple to create, integrate, and automate data pipelines across on-premises and cloud technologies. It is very comprehensive.

What is most valuable?

Control-M provides workflow orchestration, including scheduling, deploying, managing, and monitoring workflows. It helps us meet our service-level agreements. It is integrated into our CI/CD pipeline. It enhances our operational productivity.

What needs improvement?

The UI can be challenging for new users due to its learning curve. Additionally, there are some errors during automation. More detailed logs would be helpful.

We would also like enhanced API support. The APIs should be more comprehensive.

Buyer's Guide
Control-M
July 2025
Learn what your peers think about Control-M. Get advice and tips from experienced pros sharing their opinions. Updated: July 2025.
861,524 professionals have used our research since 2012.

For how long have I used the solution?

I have been using Control-M for almost two and a half years.

What do I think about the stability of the solution?

Control-M is stable so far, with no issues regarding crashing or lagging.

What do I think about the scalability of the solution?

Control-M is very scalable. It can absorb more workload wherever needed.

How are customer service and support?

We have had occasional response issues with their customer service. They do not always provide timely support.

How would you rate customer service and support?

Neutral

Which solution did I use previously and why did I switch?

We have used Redwood RunMyJobs and Apache Airflow. Control-M offers more features under one umbrella. It has cloud-native support, real-time analytics, and other features.

How was the initial setup?

The initial setup was challenging due to network setup issues and a lack of timely support from the service team. Its implementation took about a month. We did not have any downtime.

What about the implementation team?

We did the deployment in-house without using an integrator or consultant.

What's my experience with pricing, setup cost, and licensing?

Pricing is generally affordable, though some features cost a bit more.

What other advice do I have?

New users should familiarize themselves with the tools and undergo training. It is essential to understand the necessity of using Control-M in your organization. I would recommend starting with your workflows and gradually integrating it with all the tools.

I would rate Control-M an eight out of ten.

Which deployment model are you using for this solution?

Public Cloud
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
Flag as inappropriate
PeerSpot user
Pedro Fuentes - PeerSpot reviewer
System Engineer at Community Loans of America, Inc.
Real User
Top 10
Cost-effective, excellent support, and centralized access and control
Pros and Cons
  • "In Helix Control-M, we have the automation API that allows us to customize and do integrations easily in any script, such as Java or Python. It is all integrated within the integration API."
  • "I talked to Control-M guys back in October or November when they had a gathering here in Atlanta. We talked about not being able to go back in history in Helix Control-M for more than two weeks. We submitted a request for enhancement. They told us that they are working on it, and they are thinking of expanding that to 30 days. We would like to see it expand to 90 days, but they are working on it."

What is our primary use case?

Control-M is a job scheduler. You can schedule FTP jobs or use scripts within Control-M. You can also execute commands when necessary to schedule, or you can just run a script that is hosted on a server. Based on the schedule, you can orchestrate or automate jobs. You can set dependencies between jobs. You can correlate and create a sequence of your jobs and execute them in the order you wish. You can set the variables and options that you like. You can set the prerequisites and post-job activities after the completion, such as reports analysis, emails, etc.

How has it helped my organization?

Helix Control-M is critical for us. If we do not have a job scheduler like this, we will have to have people running 366 different jobs on a daily basis and 24 hours a day. These are the jobs that we run from midnight to midnight every single day on a scheduled basis.

I do not use Helix Control-M's Python client or Airflow. I am using the web client. I do not create jobs. I give my users access to create their own jobs. I just maintain the agents and keep the administration going. When they have questions, they come up to me and ask. We just use the web interface to go into the planning mode and create our jobs, folders, dependencies, etc.

Control-M has helped to give business users visibility and control over their jobs. Both Control-M and Helix Control-M allow me to give users control over their own set of jobs. They can log in and orchestrate their jobs as they want. They can also troubleshoot them on their own. It makes that easy. I just have to be hands-off and stand by in case they need assistance, but once the tool is deployed and every agent is up and running, it is easy. The people who have the jobs running or the job scheduled know about their own jobs. They know their own demands. They have control over the decision of when they are going to run it and how they are going to run it. It makes it very simple, and it helps.

You can set up your users and define whether they have admin privileges or they can just affect a set of jobs.

The fact that it is all centralized in the web browser makes it easily accessible from everywhere. All my users are IT people. They do different things. They do databases. They do informatics. They do development and things of that nature. To business users, such as board members of the company, we can give them reports on, for example, how the business closed and how much profits were there, or if all the transactions were submitted to the bank on time. If not, what were they missing? We can provide all things of that nature. We can pull it all up in a report and then schedule it on a daily basis or weekly basis. It is simple.

There have been cost savings with Helix Control-M. The license that the company was paying for Control-M, including support, was three grand more expensive than Helix Control-M. With Control-M, we also had to have an administrator dedicated to maintaining Control-M on-premises. That cost of having a person dedicated to doing just Control-M jobs is gone after we moved to the cloud. We are not only getting more money in savings; we are also making a better distribution and use of our time. By not having a dedicated person, we are saving a couple of grand. We are saving on the license and, of course, resources. We do not have to have dedicated resources such as servers. We do it virtually. We do not need to have resources reserved for the server and database. I just have to deploy the agent, which can run multiple instances in my cluster. They can share resources, which is another saving there.

What is most valuable?

In Helix Control-M, we have the automation API that allows us to customize and do integrations easily in any script, such as Java or Python. It is all integrated within the integration API.

Every year, they add another set of automation or compatibility with different applications. They are capable of integrating with Informatica, AWS, etc. You can schedule queries directly from Control-M on databases without having a server or agent. You can do scan jobs or queries directly. Every other month, they are doing releases, and they have tons of new integrations, which makes it compatible with more and more applications around the cloud.

What needs improvement?

They have a department that handles requests for enhancements. I talked to Control-M guys back in October or November when they had a gathering here in Atlanta. We talked about not being able to go back in history in Helix Control-M for more than two weeks. We submitted a request for enhancement. They told us that they are working on it, and they are thinking of expanding that to 30 days. We would like to see it expand to 90 days, but they are working on it. In Control-M, we were able to go back 180 days, but that was on-prem. The storage of that data was on our own servers. We know that storage is money, and we do not expect them to store that much of the data, but at least 30 to 60 days seem proper.

For how long have I used the solution?

Community Loans of America has been using Control-M since version 6. It has been at least 10 years.

What do I think about the stability of the solution?

I have not had any downtime with Helix Control-M. All the upgrades are scheduled, and they give us a time window when they think they are going to schedule them, and we adjust. I have not seen anybody notice it. The jobs get held before the update, and they start automatically after the update. If anybody noticed it, that was because I had to tell them that a maintenance window was coming up and to be aware of it.

How are customer service and support?

I contacted their support a couple of times to ask them about an error that I did not understand. They have three guys who are pretty handy. When you ask questions, you, of course, have to wait at least 24 hours for someone to respond. They are pretty fast. Most of my queries were responded to within the same day, which is great. I would, for sure, rate their support a solid 10 out of 10.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We were just using Control-M. We did not have any other scheduler. We migrated from Control-M to Helix Control-M. 

Control-M is on-premises, and it requires a dedicated administrator. Control-M has three major pieces. It has a Control-M server. It has a main agent, and it has a database. If you have HA, you will have the same things at a secondary location, so you will have to manage the cluster and make sure that all the pieces are working together. If, for some reason, one side fails, HA tries to recover in the second location. The management or the administration side of things is a challenge. It requires a dedicated person. Our main Control-M guy left us six years ago. Every time we had an issue with Control-M, it took us three or four hours to put it back where it should be. By migrating to Helix Control-M, our biggest success was getting away from the administration. Having Helix Control-M, which is a cloud product, allows us to use all the advantages of the job scheduler without handling the administration of our own servers.

If I compare Helix Control-M with what I had to do on Control-M on-prem, the process is very similar. The calendar has changed though. There was an advantage with Control-M that you could specify when was your new day load. Our new day load was every day at 9 AM in the morning. With Helix Control-M, we have to have only midnight as a new load because of the change of the date. It was a big challenge because we had to reorchestrate all the jobs to suit the new day load being moved from 9 AM to midnight. 

Essentially, scheduling a job or creating a new job requires the same effort in both applications. The advantage of Helix Control-M is that I do not depend on a single agent to pull FTP profiles. All of them are centralized. It does not matter which agent I am using. I have access to the whole list. In Control-M, FTP profiles had to be added to the agents that were being used. Helix Control-M has made it easy to orchestrate data pipelines in production because now, I do not have to worry about the whole backend of Control-M. I am sure that it is up to date, and I can log in reliably, load jobs, and orchestrate them as I need.

I once tried to migrate Control-M to something else called RunMyJobs. Compared to RunMyJobs, I would definitely go for Helix Control-M.

How was the initial setup?

In terms of our environment, we are a mixed shop. The majority of our products are on-prem. We have a Nutanix cluster in our data center, and that is where we host the majority of our things. We have maybe one or two devices on AWS. For Azure, I know that we have a license because it comes with our enterprise Microsoft 365 license, but I do not recall having any hosting there.

For migration, they have a migration tool that makes it very easy. You can run this migration tool, and it will export all your current jobs in a JSON file. It will try to import them on the tenant in the cloud on Helix Control-M. We faced a few challenges here and there because at the time we did it, some features were missing in Helix Control-M or were not supported, but they were supported in Control-M. We used to have dual endpoint profiles for the MSP file transfer or the ASP. FTP jobs have profiles where the server address, user password, or key gets stored. In Control-M, you could have a single profile with two endpoints. You could have Host A and Host B in a single profile. That was not supported with Helix Control-M at the time we decided to migrate, so we had the challenge of converting all those dual endpoint profiles into single endpoint profiles to be able to be imported. I know for a fact now that it is no longer an issue because they now allow you to create dual endpoint profiles in Helix Control-M, but it was a challenge at the time.

Fixing things here and there and making it compatible took about six months. Those six months were not just because of how hard it was to migrate. It was a combination of the challenges of migration and other tasks that we have not been doing because we could not afford to have a person dedicated entirely only to Control-M. Effectively, the time dedicated exclusively to the migration was two and a half months, but the migration was distributed in a six-month calendar because of other duties and tasks that I had to perform.

What about the implementation team?

We got help from VPMA. VPMA is our reseller. We purchased a license of Control-M through VPMA, and they have support and all that. We get help from them. They helped us to run the Orchestrator or the migration tool from BMC. They told us where the odd points were, and then we went to do it on our own. We came back and reviewed them again and kept fixing them.

Overall, we had three people. We had one person from VPMA, and then there was me and one of my technicians to assist me.

What's my experience with pricing, setup cost, and licensing?

It is not bad. The company can afford it, and it pays for itself. We have those jobs running automatically.

What other advice do I have?

I would rate Helix Control-M a 10 out of 10. I like Helix Control-M. 

Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Buyer's Guide
Control-M
July 2025
Learn what your peers think about Control-M. Get advice and tips from experienced pros sharing their opinions. Updated: July 2025.
861,524 professionals have used our research since 2012.
Patrick Byrne - PeerSpot reviewer
Higher Executive Officer ICT at Irish Government
Real User
Top 20
Provides batch management and reduced the need for manual intervention
Pros and Cons
  • "It's very easy to use. Compared to other softwares, Control-M has significantly simplified our monthly release process, making it easier to move things forward."
  • "There are numerous boxes to tick and things to check to ensure everything is in order before the upgrade happens. The process is very long"

What is our primary use case?

We use Control-M for batch automation. Previously, all of our batch work was manual, but now Control-M has significantly reduced the need for manual intervention. As a result, our batch processes are now 99% automated.

How has it helped my organization?

It's so easy to navigate, and especially for new hires, it's very straightforward to show them around the client because it is user-friendly. It's very easy to use. Compared to other softwares, Control-M has significantly simplified our monthly release process, making it easier to move things forward.

What needs improvement?

We're upgrading Control-M, and the process is very long. There are numerous boxes to tick and things to check to ensure everything is in order before the upgrade happens. We run three instances of Control-M, and making various changes for each is challenging.

For how long have I used the solution?

I have been using Control-M for five years.

What do I think about the stability of the solution?

You might experience a brief connection issue, but it usually resolves within a few minutes. The problem is related to the web server.

What do I think about the scalability of the solution?

Scalability is excellent. We utilize only about 20% of Control-M's capabilities. 

How are customer service and support?

Support is helpful, and the online community is very good. There's the community forum, which I use regularly to find answers to questions. BMC has been very helpful in that space. They were extremely fast and solved a difficult problem our in-house team couldn't solve in a matter of minutes. 

How would you rate customer service and support?

Positive

How was the initial setup?

The initial setup is straightforward. We used to use in-house software.

We have three different environments where people can work. People can use our development instance of Control-M to work on their batch processes before they go live, allowing them to experiment and refine until they get it right.

What other advice do I have?

It's much simpler now. Everything was a manual batch job. Using the features of Control-M every day makes our batch processing so much easier.

It makes our lives so much easier. For our operations team, which runs our daily batch overnight, viewing everything as it happens has been an absolute lifesaver, especially if things go wrong overnight. It's great to have that visibility. It has also sped up our process, reducing overhead and weekend overtime. Batch processing is much quicker now, resulting in fewer manual errors.

Control-M has so much functionality that even if you initially purchase it to handle a specific part of your batch work, it can offer much more. We've progressed beyond traditional batch processing to include MFT, which has been incredibly useful. Our file watchers and other automation features have significantly simplified our workflows and made our lives much easier.

Overall, I rate the solution a ten out of ten.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
IT MSP at Ryerson
Real User
Creates cost-efficiencies, saves time on scheduling and data efficiency, and provides better data management
Pros and Cons
  • "The solution is innovative. Specifically for the overseas and time differences, you can feel the efficiency of Batch Impact Manager on jobs, batch processing, and impact management. It works the best on these kinds of issues. It saves us time and money, which is important. We save a lot using Control-M."
  • "I am unsure if Control-M is compliant with Microsoft Azure environment integrations. We have some clients in Azure environments. Specifically, in Canada, government agencies and nonprofits mostly use Microsoft Azure."

What is our primary use case?

We are using it for job scheduling, shift scheduling, etc. It is pretty much orchestrating all the job shifts for the IT team or core team.

We specialize in security, which means 24/7 your system or team needs to be ready for anything or anybody in the world, independent of even time differences. If you are managing your client's services from Europe, or anywhere else in the world, Control-M makes it easier to do scheduling, saving people time.

Since we have different satellite offices in Toronto and Ottawa, we use different role accessibility in different locations. That is why we are using it on-premise. However, in the next six months or so, we are planning to go to a hybrid cloud environment on Control-M since we are adding two or more satellite offices. We think that it will be more manageable if we implement it in a hybrid cloud environment.

How has it helped my organization?

The solution is innovative. Specifically for the overseas and time differences, you can feel the efficiency of Batch Impact Manager on jobs, batch processing, and impact management. It works the best on these kinds of issues. It saves us time and money, which is important. We save a lot using Control-M.

The most important thing is it is easy to manage conversions and stuff. It is easy to convert different systems, like AWS, which saves time.

We are working with vendors, partners, and clients to manage GDPR and data privacy. This solution is good with data privacy because BMC is GDPR compliant. That is very important, especially for overseas clients and businesses.

Overall, Control-M is quite critical for our business. I would rate this as nine out of 10.

What is most valuable?

  • The monitoring
  • Workflows
  • Production
  • Scheduling the shifts
  • Timeframes for specific roles and management

Time differences are important because we have some overseas clients. That is why we are using Batch Impact Manager for the Control-M, which is very helpful for us. It detects potential blocks in advance, delays, and errors. That helps us to optimize the scheduling, then the batch workload processing as well. 

It is pretty easy to create, integrate, and automate data pipelines. It is user-friendly, not rocket science. That is what I like about the Control-M, and specifically Batch Impact Manager. You will need some orientation and need to know what you are doing if you are integrating your system, and this solution makes it easier.

We use Control-M Python Client and cloud data service integrations with GCP, which is pretty hassle-free. There aren't any problems or compliance issues. It is pretty easy to retrieve the data, do conversions, etc. They are on-time, and there is not much of a delay. 

The engineers on our team say that it is pretty easy to build, test and promote data workflows with the data coding language integrated into Control-M through the Control-M automation API. The ease of integration is eight out of 10. Python is the main language that our database managers and data engineers are using along with some other languages. 

The Control-M interface is user-friendly and easy to use. Orientation-wise, it is easy for data engineers to adapt.

What needs improvement?

Ingesting and processing data from different platforms can be a challenge. Control-M does allow integration for this with other systems to make this easier. For example, we integrate Control-M with an in-house system to do this.

I am unsure if Control-M is compliant with Microsoft Azure environment integrations. We have some clients in Azure environments. Specifically, in Canada, government agencies and nonprofits mostly use Microsoft Azure.

For how long have I used the solution?

I have been using it for almost a year and a half.

What do I think about the stability of the solution?

The stability is very good. My impressions of the stability are very positive.

What do I think about the scalability of the solution?

I would rate the scalability as nine out of 10.

How are customer service and support?

The technical support is near perfect. I would rate them as nine and a half out of 10.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We were previously using an in-house solution, but we weren't improving it much, which is why we switched to Control-M.

What was our ROI?

Control-M creates cost-efficiencies, saves time on scheduling and data efficiency, and provides better data management. We use the managed services as well because we partner with some clients at MSPs and MSSPs. This solution is also good for their environment because it is easy to access, retrieve, and work with actionable data as well as all the procedures and processes. It is good and works. I would rate it as eight out of 10. 

The total cost of ownership is impacted by the Control-M pricing as well as the overall cost of the procedures and daily batch processing. We can easily see that at the end of the year, in terms of big time and money savings.

What's my experience with pricing, setup cost, and licensing?

The pricing and licensing could be better. However, when I compare Control-M pricing with JAMS, Control-M is still better priced than JAMS enterprise.

Which other solutions did I evaluate?

We also evaluated JAMS Scheduler, which is also a workload automation solution. The pricing for Control-M was better and has good predictive maintenance that is better than JAMS. Control-M is also more integrated with Google for different solutions.

What other advice do I have?

Control-M is better for the cloud. Specifically, the hybrid cloud is the best. On-premise is still okay, but it depends. Its hybrid cloud environment works better and is optimized in a better way to save money and time. Its implementation is easy from the cloud GCP and AWS. Microsoft Azure is not there yet, but otherwise, it is perfect.

I would rate the solution as nine out of 10.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
reviewer1631958 - PeerSpot reviewer
Maintenance Manager at a transportation company with 10,001+ employees
Real User
We have seen quicker file transfers with more visibility and stability
Pros and Cons
  • "If they have ad hoc requirements, then they can theoretically schedule their own file transfers with the Self Service. We are trying to push as much work back to the customers or developers that have that requirement, because they prefer to help themselves, if possible. We try empowering them and enabling them through Control-M, especially for file transfers, because it is a much broader base of the business then just with batch scheduling. Typically, with SAP batch scheduling, it would work with dedicated teams. With file transfers, the entire business is involved. There are business users, end users, etc. It definitely needs to be as simple as possible and as managed as well as possible. They need to manage it themselves, if possible, because our team is not growing but the number of customers, applications, and jobs are growing. We need to hand back some of the responsibility to the customer for them to resolve and action it."
  • "The high availability that comes from BMC with its supplied Postgres database is very limited. Even using your customer-supplied Postgres database is problematic. We have engaged with them regarding this, but it is difficult. My company doesn't want to do this and BMC doesn't want to do that. We just need to find some middle ground to get the proper high availability. We're also moving away, like the rest of the world, from the more expensive offerings, like Oracle. We are trying to use Postgres, which is free. The stability is good. It is just that the high availability configuration is not ideal. It could be better."

What is our primary use case?

We schedule the majority of our SAP jobs Control-M. We do that globally for all the production plants. We have tens of thousands of SAP jobs and managed file transfer.

SAP batch and managed file transfer are critical processes that we have automated. We are in the process of replacing Connect:Direct and SecureTransport, the legacy file transfer solution, with Managed File Transfer (MFT). That is on the global scale. 

The Control-M for Informatica is gaining a lot of popularity, primarily in the financial side of the business. They have a lot of security restrictions that make their jobs very difficult. Also, there are cost issues for Informatica, e.g., anytime they execute a workflow in Informatica, they get billed for it. We are adapting the solution to not scrum the workflow every half an hour or hour because they pay for it, but only when it is needed. Therefore, we can do a database query and check if there are new records that need to be processed. If there are no records to be processed, then depending on that output, we either run the Informatica job or leave it and check again for maybe half an hour. We are optimizing, saving money for the customers and ourselves, while reducing the number of executions, jobs, etc.

We are using on-premises. We have been for many years. We are aware of the new Helix offering, which is a SaaS/cloud offering from BMC, but it is not really ready for enterprise yet, not at our scale. We are doing some cloud, though not the Helix offering. I have installations in the cloud using Azure and AWS. We are not fully functioning there yet. We are waiting for the demand, but we are aware of the cloud opportunities and making use of them.

We have been busy upgrading to version 9.0.20 Fix Pack 100 but our production environment is still on 9.0.19 Fix Pack 200.

How has it helped my organization?

We use Control-M as part of our DevOps automation toolchains and leverage its “as-code” interfaces for developers. We have found that a lot of the new customers who are developing for cloud prefer to use the API and would like to test for themselves. That is really where Jobs-as-Code comes in. They can test and fail quickly the agile way. We definitely have some customers who are using that.

We have seen quicker file transfers with more visibility and stability. Because data transfers are part of the Control-M tool, they form as part of the normal workflow. We see the value in that.

If they have ad hoc requirements, then they can theoretically schedule their own file transfers with the Self Service. We are trying to push as much work back to the customers or developers that have that requirement, because they prefer to help themselves, if possible. We try empowering them and enabling them through Control-M, especially for file transfers, because it is a much broader base of the business then just with batch scheduling. Typically, with SAP batch scheduling, it would work with dedicated teams. With file transfers, the entire business is involved. There are business users, end users, etc. It definitely needs to be as simple as possible and as managed as well as possible. They need to manage it themselves, if possible, because our team is not growing but the number of customers, applications, and jobs are growing. We need to hand back some of the responsibility to the customer for them to resolve and action it.

What is most valuable?

A new feature, which we deployed about two years ago, is the Managed File Transfer (MFT). We also use Managed File Transfer Enterprise (MFTE) for external transfers of our biggest use cases. 

Another valuable feature would definitely be the MFT dashboard that is now available in Control-M natively. It is easy to just search for jobs, files, etc. Instead of the customers contacting us to find out what happened, when it happened, and why it happened, they are able to service themselves. This allows us to cut down on operational staff, costs, and time because customers can manage it themselves to a degree.

The most valuable feature is definitely the Self Service. A couple of years ago, it was available, but not with the features that it is today. There wasn't really uptake on it, although it was available. We have seen a steady growth in the number of users using it and what they are using it to do. They are using Self Service to schedule by themselves and do monitoring by themselves. They interact with their schedules. Also, the performance of Self Service is very user-friendly and more accessible. That is one of the features that we use a lot lately.

The reporting has definitely improved over the years. We are definitely doing more of that as well. We are definitely seeing more value in reporting on the batch schedules, optimizing it and seeing if we can cut costs. 

What needs improvement?

The reporting has improved. It is not where it should be yet, but we have seen improvements. The biggest thing for me is the restrictions regarding templates for reporting. You can't create your report with your own parameters. We have a meeting weekly with BMC and our customer lifecycle architect, and this comes up quite frequently. We have been privileged enough to do work with the developers. They are aware of the requirements regarding reporting and what our customers are asking for.

What I found lately about the YouTube videos, specifically, is that they are very simple. Usually, when I watch a video, I would read the manual, instructions, etc. to see if I understand it. I would hope that the interactive sessions, Q&As, or videos could be used to handle more complex issues of what they're discussing. An example would be the LDAP authentication for the Enterprise Manager. They would typically just go through the steps that are in the documentation. What people typically looking at those videos are looking for is how to do the more complex setup, doing it with SSL and distributed Active Directory data mines. Things that are not documented. I find those videos helpful for somebody who is too lazy to read the manual. I expect them to handle more than what is available in the documentation and the more complex situations.

The high availability that comes from BMC with its supplied Postgres database is very limited. Even using your customer-supplied Postgres database is problematic. We have engaged with them regarding this, but it is difficult. My company doesn't want to do this and BMC doesn't want to do that. We just need to find some middle ground to get the proper high availability.
We're also moving away, like the rest of the world, from the more expensive offerings, like Oracle. We are trying to use Postgres, which is free. The stability is good. It is just that the high availability configuration is not ideal. It could be better.

For how long have I used the solution?

I have been using Control-M for 12 years.

What do I think about the stability of the solution?

Control-M is really stable. We have seen that throughout the years. I have had customers who have been running version 6.3 for seven years after support stopped. It has been running for three years straight, without a reboot or restart, doing its job. We have actually had issues with customers who don't want to upgrade. They have said, "This stuff is working perfectly. Just leave it alone because it just doesn't go down." 

We have a saying in our department as well. When somebody says there is a problem, we say, "It's not Control-M. Check everything else. Check the server, network, and database. It's not Control-M." 99 out of 100 times, we are right. It is either infrastructure or something else, but it is not Control-M.

What do I think about the scalability of the solution?

I have never run into any problems scaling, either vertically or horizontally, with Control-M. In each version, it just gets better. I am really happy with that.

We were one of probably the first companies who bought MFTE, and it was not ready yet. It didn't scale properly. It didn't offer the functionality that the competing tools that we were currently using had. It's grown tremendously because of our input and feedback directly to the developers and BMC. I'm not complaining about it, but it put us back a bit. We have learned not to be a very early adopter. We have seen the same with the cloud. Everybody wants to jump on the cloud, but nobody knows why. They just want to do Cloud. We've made a substantial investment with MFTE. It was a couple of hundred thousand euros, and it was not ready yet for our enterprise requirements.

Our monitoring team who does 24/7 monitoring. They handle the alerts. They check their job flows. They make sure escalations are going through. If tickets need to be logged, make sure that gets done. They also interact with ad hoc requests from customers. 

There is the scheduling team who does the job definitions, updates, etc. 

There is the administration team, which I'm part of, with administrators who look after the infrastructure, Enterprise Manager, servers, agents, gateways, etc. Recently, we also have a dedicated MFT team that only looks after MFT because of the huge number of customers, requests, and requirements.

Other customers who use it are really all across the board. We had a presentation last week to our bigger department that is worldwide, but which we are a part of in South Africa. We have noticed about 52 main departments, then the sub-departments, between them. A lot of them sit right across the enterprise. Typically, the most active users would be SAP users who checks for output on the jobs running on Control-M. It is just 10 times easier to do it in Control-M than on SAP itself. We also manage to keep the output for longer than SAP. What they can't find on SAP after seven or 14 days, they can usually find with us, e.g., outputs for the jobs or logs. 

There are the MFT users who love being able to see each morning that their file was transferred, how long it took, and how big the file was. A lot of self-service users are using the Self Service function. Team leads and operational staff use it most.

How are customer service and technical support?

I love support and the support people. It is very good. Because we are quite a mature customer and the whole team has a lot of experience (sometimes more than the support people), if they don't realize the seriousness of the situation, then we would not escalate but just to make our customer lifecycle architect aware by saying, "We are not feeling this case is getting the required personnel on it. We need somebody more senior. We don't have time to cover the basics that the first line support is trying to deal with. We've been over that." Overall, I would rate the technical support as nine out of 10.

Which solution did I use previously and why did I switch?

Previously, we used a big SAP solution, which was not a commercial, and specifically designed for our company.

We have recently taken over a mainframe migration as well as the scheduling was on TWS, which is IBM's scheduling software on the mainframe z/OS. We moved that all over to Control-M. That was a combination of SAP jobs, Informatica jobs, database jobs, and normal script jobs. So, we use a bit of everything. We have also used the automation API a lot for interfacing with Control-M and other middleware tools, but primarily it is SAP and file transfer.

We use Control-M to integrate file transfers within our application workflows. It integrates with the tools that we are replacing, i.e., Connect:Direct, which is quite a legacy tool, and our old IBM tool, which we have been using for more than 15 years and has no visibility. With Control-M, you get visibility on your file transfers and how it mostly interacts with your batch schedule. Something gets created, it's sent over, and then it gets processed. Control-M has already been part of the executing, extracting, import, or processing. Now, with the file transfer, customers can see the entire workflow from the data being generated, transferred, and processed. This resolves a lot of complexities because you used to need to contact three different teams to find out if the file arrived and was processed. One tool does all of that now.

There isn't a lot of new functionality that our previous tools didn't have. It is just re-consolidating all the tools that we need into a single one. That makes it much simpler. There is one team to contact globally for file transfers, and that makes it easy. It provides visibility with its Self Service that wasn't available with Connect:Direct or SecureTransport. Our customers are quite happy to have that. We can also provide reports. 

SecureTransport competes with MFTE. There isn't a conversion tool for that yet. Connect:Direct simply provides the means for a conversion tool, but it gets integrated into scripts and applications. It's very difficult to migrate or extract that data.

How was the initial setup?

The initial setup is straightforward. It changed a lot over the years as well, but in the nicest way. You have minimal downtime with the upgrades on Enterprise Manager as well as the Control-M servers. A lot of preparation is done before the tool is shut down for the upgrade. Our downtime used to be at least an hour for upgrades or migrations. That has typically come down to 10, 15, or 20 minutes, depending on the size of the server. It is definitely more stable and understandable.

We have also noticed that the exception handling is much better if there are issues. We don't get that many surprises. The errors are understandable. The agent upgrades have zero downtime, so that is just amazing. All the patching and maintenance is centralized. We have migrated our development and integration environments to 9.0.20 in the last month or two. That went very smoothly. We will start with production next week. We have been through this quite a number of times. We came from version 7 to version 9 to versions 9.0.19 and 9.0.20. We do all the upgrades in-house.

What about the implementation team?

We do it all ourselves. If we get stuck, we would contact BMC. At my previous job, we were a partner for BMC in South Africa, and I was on the support side for BMC. It is only we need to open tickets for bugs or problems that we contact BMC. Typically, upgrades and migrations, we handle those in-house.

There are three people full-time on the administrative side. We have a global setup: Europe, Mexico, America, Africa, and China. We have tons of virtual machines and hundreds and hundreds of agents, and even more that we might host.

What was our ROI?

I know we have already budgeted for more tasks. The company is very happy with the performance of our teams, specifically the South African team. We are really doing more with good tools and less people. There is definitely a return on investment, just from the stability and visibility which has improved a lot.

On the effort side, we have definitely seen a lot of savings. We have some bigger projects that are automating the schedule and removing human intervention. These have reduced department staff/headcount, by about 50%, when we were able to automate the batch side of it, because also our department offers monitoring and operations as part of our service. We have a dedicated monitoring team. Whatever runs in Control-M, that is monitored by us and escalated, if needed. 

Departments now have multiple scheduling tools between the mainframe, distributed systems, and cloud. Control-M brings all of that, e.g., we have it on a single pane of glass so we can see the exact execution on the mainframe, the execution on the line, and the execution in the cloud. This is instead of using three or four different tools. Therefore, the complexity of batch monitoring and scheduling has decreased as well with the standardization of Control-M. That is definitely one of the big advantages that we have seen.

What's my experience with pricing, setup cost, and licensing?

It is expensive. We have a lot of customers who complained initially about the costs. Because it's not just the licensing, unfortunately. It's the infrastructure, salaries, etc. I like the licensing model. It is pretty straightforward. We are on the task license. I know that we have some really good discounts. Our BMC account manager makes sure that we stay below the license count as well as checking for growth. Overall, it's good. The licensing is simple enough for me. It is a bit expensive. Especially with the cloud coming in, we might see the licensing change in the future, but I'm guessing.

This is now from my previous years as support for banks and big companies. If it's not enterprise scale, I find that it's too expensive for smaller companies. You really have to be quite big and need to have a dedicated support staff to run it, then you'll be fine. What we've seen at smaller companies, it's too expensive because they want to automate everything. Now, stuff that can literally run once a day for the rest of their lives is costing them $3 a job a day. It becomes too expensive, eventually. They are not seeing the return on investment because it's not business critical. Nobody is going to die or they're going to lose money if that job didn't run exactly at 11 minutes past 4:00. It's definitely for bigger enterprise companies, especially banks or healthcare providers. We have had an instance where Control-M was unavailable due to external factors for 20 minutes and there was a loss of almost a million euros because the solution involved logistics. 

Which other solutions did I evaluate?

We have done the usual crontab migration. Everything is in crontab or Windows Scheduler. Typically, we end up with a migration, even if it's from a known tool, where we end by exporting it into Excel and converting it into job definitions with a script. We have been involved in that, but nothing using BMC tools.

When I joined the company, I first supported them through the local partner. Because we have such a vast array of scheduling tools, they went through a PoC and business case. We evaluated three or four tools, where BMC Control-M was one. Quite soon, because the company was already using Control-M in Africa and China, they were looking for global solutions to see if it really could create change.  

What it came down to was ease of use, enterprise capability, and BMC was already in the company with ITSM and a couple of other products as well. They had a good relationship with us. We consulted with other customers who have used it as well as references because it was expensive. It was definitely the most expensive solution then, out of the four. However, we didn't want to go five years down the line and then have to change again because of issues.

What other advice do I have?

We have had a very good run with Control-M. I love it.

With the move to big data and especially with our AWS Cloud presence, we have a data lake. We are in discussions with the analytics teams about how they can utilize Control-M in the cloud for analytics, big data, etc. However, at the moment, it is not a big deal.

What we have found with the Jobs-as-Code is that customers need to understand Control-M better, how the scheduling works, the knowledge around it, its conditions, etc. It took some time for the developers to get used to Control-M, then Jobs-as-Code. They are now confident with it. We are presenting twice weekly. We have an open forum for interested parties about Control-M or our department, Enterprise Scheduling and File Transfer, where we have a dedicated session about Jobs-as-Code. If there are questions about how other departments are doing it, if there is a better way to do it, if they are able to save on the number of jobs, can we make them rerun, or instead of creating 10 jobs, can it be done with five jobs? So, there is not a lot going from Jobs-as-Code directly into production, but we have a couple of parties, especially on the cloud front, who are very interested in it.

The solution is enterprise scale. Also, if you want to integrate all your applications into one view and offer all the functionality across the board, such as file transfer, scheduling, cloud, and on-prem, then you can create your own application integrations to integrate with applications that's not supported currently by BMC, like APIs. For top 100 enterprises, there isn't another better tool on the market for enterprise.

I would rate it as a nine out of 10.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
reviewer1657833 - PeerSpot reviewer
AVP - Systems Engineer at a financial services firm with 10,001+ employees
Real User
Allows us to integrate file transfers more readily, resolve issues quickly, and orchestrate a diverse landscape of vendor products
Pros and Cons
  • "The File Transfer component is quite valuable. The integration with products such as Informatica and SAP are very valuable to us as well. Rather than having to build our own interface into those products, we can use the ones that come out of the box. The integration with databases is valuable as well. We use database jobs quite a bit."
  • "A lot of the areas of improvement revolve around Automation API because that area is constantly evolving. It is constantly changing, and it is constantly being updated. There are some bugs that are introduced from one version to the next. So, the regression testing doesn't seem to capture some of the bugs that have been fixed in prior versions, and those bugs are then reintroduced in later versions."

What is our primary use case?

Control-M supports a lot of business processes. It supports some of the HR functions. I don't know if payroll is directly supported, but we do run jobs through PeopleSoft, which obviously impacts HR. Recently, we've started using the SAP module. So, we're making a transition from PeopleSoft to SAP, and I also see some payroll functions happening there.

How has it helped my organization?

We use Control-M to orchestrate a diverse landscape of vendor products such as Pega, MuleSoft, etc. File transfers and data feeds fetching are quite important for us. So, a lot of data processing happens through Control-M.

Control-M provides us with a unified view where we can easily define, orchestrate, and monitor all of our application workflows and data pipelines. Of course, such a diverse landscape requires you to make the effort to utilize Control-M to tie everything together or to act as the glue. Once you do that, everything is clearly defined, and you can view these disparate systems using one unified pane. If you don't define it correctly, then obviously Control-M won't have that insight, and so you'll have to go to multiple locations to go look at your job statuses.

We use its web interface. It is primarily for the application support teams to go monitor their own jobs. The jobs defined within Control-M are tightly controlled by a specific group of people. There are also people who need access to view that the jobs were completed successfully or why the jobs may have failed. These people are given access through Control-M web to view and monitor the jobs that they support or the applications they support. They're usually able to log on without having to install any client on their personal workstations. So, it's quite convenient. We have not implemented its mobile interface.

The integrated file transfers with our application workflows have certainly sped up our business service delivery by 80%. It has allowed the business to integrate file transfers more readily. Prior to utilizing the Control-M module, people had to write their own file transfer scripts in a scripting language of their choice to vary degrees of effectiveness. With the integrated File Transfer solution within Control-M, there is a standardized way of performing file transfers along with the capability of file watching and grabbing the file names that were transferred, making it much more versatile.

Control-M can immediately report when a job fails. If you have proper monitoring in place, you're notified immediately when your business flows are impacted. In the past, when you run jobs using Cron or just wrote shell scripts, you're really left in the dark because they don't necessarily report even from within Control-M. Implementing Control-M has made the business realize how critical and important it is to have proper error coding within the scripts that they schedule. If the scripts don't necessarily report any errors or redirect the system output into log files, when a job fails, there is no way to detect that.

We've automated many time-consuming business reports and other things that were very manual and took a tremendous amount of manhours. We've also automated a lot of maintenance using Control-M. We've integrated with Ansible Tower. So, we now are able to run Ansible playbooks and Ansible job templates. With the scheduling capability and the multitude of integrations that Control-M offers, it really acts as the unifying glue and as a communicator and orchestrator across the enterprise. With Ansible Tower, you can run a number of playbooks through it to perform patching and reboots and whatever maintenance that the infrastructure teams require, but you can't really do it when the business is still operating, or you can't do it when that business is operating, but you could do it for another business that's not operating at the moment. It is very hard to coordinate that without knowing which lines of business have jobs running or things like that. With Control-M, you can see that and you can actually enact workload policies to put jobs on hold prior to running Ansible playbooks. Once your Ansible playbook is complete, you can release the jobs again by deactivating the workload policies. So, it makes those processes very streamlined.

We do use the Role-Based Administration feature. We have been allowing other groups to gain more control over their agents so that they can define connection profiles, and they can do a little bit more on their side without inundating the main team with a lot of tasks. Everybody is happier. They can get things done faster, and they have immediate feedback and response because they're in control. The main Control-M team is not inundated with a lot of different requests from various teams to do a number of mechanical tasks. They don't get asked to create the connection profile for a database. People have all the information there, and they can do it themselves. They can define it in a way so that only they have access to it.

It has helped us to achieve faster issue resolution. Control-M reports on the error. It is easier to view the system output of that job. Whether it is an Informatica job, a scripted job, or a database job, it is easier to go in and view the issue and then troubleshoot from there. Most of the time, you can be running from the point of failure if the jobs aren't defined correctly. In a properly defined job, I would estimate that there is a 70% to 90% reduction in the meantime to resolution.

It has helped us by improving our service-level operations performance. We've built integration between Control-M and our ITSM, which is ServiceNow, and that has certainly allowed us to gain more visibility within our community through ServiceNow. Every time a production job fails, an incident ticket is cut, and that's highly visible. That needs to be escalated too, and there is a much more defined process to be able to resolve that issue. In the past, obviously, when you didn't have that level of visibility or that integration, there was always time lost in identifying what the issue is.

What is most valuable?

The File Transfer component is quite valuable. The integration with products such as Informatica and SAP is very valuable to us as well. Rather than having to build our own interface into those products, we can use the ones that come out of the box. The integration with databases is valuable as well. We use database jobs quite a bit. The file watcher component is also indispensable when integrating with other applications that generate files, instead of triggering a workflow based on time.

What needs improvement?

We have been experimenting with centralized connection profiles. There are some bugs to be worked out. So, we don't feel 100% comfortable with only using centralized connection profiles. We do have a mix of control on agents out there, which leads to some complications because earlier agents do not support centralized connection profiles.

A lot of the areas of improvement revolve around Automation API because that area is constantly evolving. It is constantly changing, and it is constantly being updated. There are some bugs that are introduced from one version to the next. So, the regression testing doesn't seem to capture some of the bugs that have been fixed in prior versions, and those bugs are then reintroduced in later versions. One particular example is that we were trying to use the Automation API to fetch a number of run ads users from the environment. The username had special characters and backspace characters because it was a Windows User ID. In the documentation, there is a documented workaround for that. However, that relied on two particular settings in the Tomcat web server. I later found out that these settings work out-of-the-box for version 9.0.19, but those two options were not included in the config file for 9.0.20. So, it led to a little bit of confusion and a lot of time trying to diagnose, both with support and the BMC community, what is the issue. Ultimately, we did resolve that, but that is time spent that really shouldn't have been spent. It had obviously been working in 9.0.19, and I don't know why that was missed in 9.0.20, but that's a primary example of an improvement that can happen.

We've also noticed that the Control-M agents themselves now run Java components. Over time, they tend to destabilize. It could be because garbage collection isn't happening, or something is not happening. We then realize that the agent is consuming quite a large amount of memory resources on the servers themselves. After recycling the agents and releasing that memory, things go back to normal, but there are times when the agent becomes unresponsive. The jobs get submitted, and nothing executes, but we don't know about it until somebody says, "Hey, but my job isn't running." When we look at it, it says Executing within the GUI, but there is no actual process running on the server. So, there is some disconnect there. There is no alerting function or the agent there that says, "Hey, I'm not responding." It is not showing up in the x alerts or anything like that.

The integrated guides have not been that helpful to us. I do find a lot of the how-to videos on the knowledge portal to be useful. However, there are some videos where the directions don't always match with some of the implementations. There are some typos here and there, but overall, those have been more helpful for us.

Its pricing and licensing could be a little bit better.  The regular Managed File Transfer piece, is a little overpriced, especially for folks who already have licensed Advanced File Transfer.

What I'm also noticing when I'm trying to recruit for Control-M positions is that the talent pool is quite small. There's not a whole lot of companies that utilize Control-M, and if they do, most people don't want to let their Control-M resources go if they're good. There is a high barrier of entry for most people to learn Control-M. There are Workbench, Automation API, and so forth mainly for developers to learn, but there are not a whole lot of resources out there for people to get more familiar with administering Control-M or things like that in terms of the technology or even awareness. So, it becomes very challenging to acquire new resources for that. A lot of the newer people coming out of college don't even know what is Control-M. If they do, they think of it as a batch scheduler, which is certainly not true in its current transformation.

Control-M is a very powerful enterprise tool, but the overall perception has not changed in the past five to six years that I've been working with Control-M. There's not much incentive for people to dive into that world. It is a very small community, and overall, the value of Control-M is not being showcased adequately, maybe at the C-level for corporations. I've had multiple conversations with other people and other companies who have already exit using Control-M. About 70% of the companies out there do not take full advantage of the capabilities in Control-M. So, that type of utilization really hampers and hinders the reputation of Control-M. That's because people then acquire this untrue concept that Control-M can only do X, Y, and Z, rather than the fact that Control-M can do so much more. I don't know if it needs a grassroots marketing movement or a top-down marketing movement, but this is what the perception is because that's what I'm hearing and that's what I'm seeing. For some of the challenges that I face working in Control-M, when I go back to my management and say, "Hey, I want to spend more money in this space," they're like, "Why? Can you justify it? This is what we see Control-M as it is. It's not going to bring us value in this area or that area." I have to go back and develop a new business case to say, "Hey, we need to upgrade to MFT enterprise or something like that." So, it definitely requires a lot more work convincing management in order to get all these components. In the past, we had to justify acquiring a workload change manager. We had to justify acquiring the workload archive. All of these bring benefits not only to our audit environment but also to the development environment, but the fact that we had to fight so hard to acquire these is challenging.

For how long have I used the solution?

I've been using Control-M for about eight years.

What do I think about the stability of the solution?

Version 9 was very stable. Once they started adding a lot of the newer Java components, the stability suffered. It seems to have gotten better in version 9.0.20, but that's could be my basic perception. 

We run a lot of database client jobs. There are some things that we've implemented that I understand can contribute to the agent instability. We sometimes extract a lot of database output and massage that output using other scripts. I've noticed there are certain things that you cannot do with it, or there are some things that contribute to the instability. For example, in the output scanning functionality, there certainly is a size limit. You probably don't want to scan anything too large because that's going to put a lot of resources on the environment.

In addition, there are times when the agent becomes unresponsive. The jobs get submitted, but nothing executes. There is no alerting function. These are the examples of instability that I've noticed. Overall, the main application itself, the EM, and the scheduler have been pretty stable.

What do I think about the scalability of the solution?

It is very scalable in terms of job execution. I haven't really explored scaling Control-M and the EM environment to a point where we have hundreds of users accessing it at a given time. That's because I don't have a hundred users who want to access that at a given time, but I do understand that you can distribute the web server more, and then have a load balancer to balance the load. I would think Control-M is a fairly scalable application.

In terms of its users, we have a lot of application support folks. We do have some developers who access Control-M mostly for the non-prod environments to execute and monitor their own jobs. There are some software engineers and operational engineers who are part of the application support teams that access Control-M. As for size or concurrent users, we have about 50 concurrent users at the max.

How are customer service and support?

I would probably give them a nine out of 10. For the most part, they're very helpful, but there's always an initial standard dialogue. For an issue, you have to collect from EM logs, agent logs, and so forth, and you submit that. Sometimes, we have done all the advanced work and submitted it, but they still come back and say, "Hey, we need the logs." It seems like that's a canned response without looking at the tickets.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We've been with Control-M for quite a long time. We have not been using anything else in my history with this organization. 

I have not looked at anything recently. I am aware there are other application orchestration solutions out there, but I have not felt the need to go explore those options at the time.

How was the initial setup?

If you're deploying using out-of-the-box options, the process is fairly straightforward. If there is some customization that needs to happen, then the process can be complex, and the documentation does not cover some of those complexities.

For the most part, we are standard out of the box. We have run into some performance issues where we had to, later on, go in and maybe make some modifications. For example, we had to stand up different gateways for various purposes just because one singular gateway was not enough to take the load in particular because we had installed a workload archive, and that was just taking up a lot of resources. Other human users were not able to perform their actions because the archive user was consuming so much of the server's resources. So, there was a lot of tweaking there, and we had to basically break out and distribute some of the components.

In terms of implementation strategy or deployment plan for Control-M, the environment always had Control-M, and we just had to upgrade the Control-M environment. We've had Control-M in our environment for quite a long time, probably when it was still version 6. So, as we progressed through different versions, we obviously had to expand the environment and the platforms. We initially started off with Control-M on AIX, and we later moved to Control-M on Linux. As you go to Linux, obviously, there is planning for high availability and production environments, disaster recovery environments, and so forth. So, you have to plan for marrying a lot of the BMC Control-M components and identifying where a load balancer may be required, or DNS ALIAS is required so that you can quickly flip over in the event something happens. Then, of course, there is sizing for the environment in terms of how many jobs are running, how many executions are happening, and so forth. This is how we plan.

What about the implementation team?

We've used the AMIGO program, and then we've performed the upgrades ourselves.

For its day-to-day administration, we have a team of five people. They're administrators and schedulers.

What was our ROI?

Its return on investment is quite high, and that's mostly because we use so many of Control-M's capabilities. We also extend those capabilities. We write our own scripts to be able to integrate Control-M with so many other applications such as Automation Anywhere, Alteryx. We have also done vice versa. We have helped other teams develop their capabilities in integrating with the REST API and Control-M. So, the ROI is quite high for our use case, but based on the conversation with some of the community partners out there, their ROI is probably quite low because they're not making use of all these new features. I don't know if it is because they don't have the skillset to make use of these new features, or their management structure or process structure is hampering them from going out there. A lot of large companies I know like to maintain the status quo, and that's why they're slow to adapt and slow to move, which is going to hurt them in the long run, but in the meantime, it can hurt the adoption of Control-M as well.

What's my experience with pricing, setup cost, and licensing?

Its pricing and licensing could be a little bit better. Based on my experience and discussions with other existing customers, everybody feels that the regular Managed File Transfer piece, not the enterprise one, is a little overpriced, especially for folks who already have licensed Advanced File Transfer. We understand that Advanced File Transfer is going away and is going to be the end of life, and there is some additional functionality built into MFT, but the additional functionality does not really correlate with the huge price increase over what we're paying for AFT already. This has actually driven a lot of people to look for alternative solutions.

I know they are now moving more towards endpoint licensing or task-based licensing. In my eyes, the value of Control-M is the ability to break down jobs from monolithic scripts. You don't want to have to wrap everything up in one monolithic script and say, "Hey, I'm executing one task because I want to save money." That defeats the purpose of controlling, and that defeats the value of Control-M. By being able to take that monolithic script and break it down into the 10 most basic components, you can monitor each step. It is self-documenting because, within Control-M, you can see how the flow will work, and you can recover from any one of those 10 steps rather than having to rerun the monolithic script should something fail. That being said, the endpoint licensing does make more sense, but maybe pricing or things like that can be more forgiving.

Which other solutions did I evaluate?

N/A

What other advice do I have?

It is worth the time and money investment to learn more about Control-M. You should learn all the features of Control-M and really explore and test out the capabilities of Control-M. That's the only way people get comfortable with what Control-M can implement. A lot of people aren't aware of just how flexible a platform Control-M is, especially with all the new features that are being added via the Automation API. These features are helping to drive Control-M and things developed in Control-M more towards a microservices model.

We are just beginning to explore using Control-M as part of our DevOps automation toolchains and leverage its “as-code” interfaces for developers. Obviously, there is a little bit of a learning curve for developers as well in order to see the value of developing Jobs-as-Code. Currently, we're walking developers through it, and we're holding their hands a little bit in terms of developing Jobs-as-Code, but we are heading in that direction because it does provide artifacts that you can version control and change quickly and easily. You can redeploy much quicker than just having the jobs defined in the graphical user interface. Previously, when you had to modify it, you either did it via the GUI, or you exported it via XML and then modified those components. Once you get the developers closer to their job flows, then you can theoretically speed up the delivery of applications along with scheduled jobs.

I don't have a whole lot of experience with other scheduling orchestration environments, but from everything that I've heard while speaking with other colleagues, I would say Control-M ranks fairly high. I would rate it a nine out of 10. Control-M usually is the platform that people are moving to, not moving away from.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Director Information Technology at a insurance company with 1,001-5,000 employees
Real User
Enabled us to consolidate and streamline our development process, while building on existing skills
Pros and Cons
  • "We used Control-M's Python Client and cloud data service integrations with AWS and, as a feature, it was very customizable. It gave us a lot of flexibility for customizing whatever data maneuver we wanted to do within a pipeline."
  • "I would like to see them adopt more cloud. Most companies don't have a single cloud, meaning we have data sources that come from different cloud providers. That may have been solved already, but supporting Azure would be an improvement because companies tend not to have only AWS and GCP."

What is our primary use case?

Our use case was mainly about consolidating our data pipeline from different sources and doing some data transformations and changes. We needed to get data from different sources into a state where we could act on it into one consolidated data set.

How has it helped my organization?

It gave us the ability to consolidate a diverse set of solutions into one comprehensive solution that streamlined our development processes. It was straightforward to adopt and we could build on existing skills without having to have 10 solutions for 10 problems.

And when it came to creating actionable data, it gave us the ability to move faster and at scale. By adopting a solution like Control-M, we were able to scale and deliver faster data transformations and maneuvers, turning data into insights in a more efficient and scalable way.

The ability to deliver faster and at scale was important. Business and management always wanted us to deliver faster and bigger and we were able to do both with the solution that we implemented using Control-M. We were able to respond faster to changes and business needs, at scale. 

Having a feature-rich solution enabled us to aggregate all of our processes into it, and that made the overall execution, from a project and portfolio perspective, a lot more efficient.

We were also able to respond to audit requests, because it's centralized, in a much more efficient way.

What is most valuable?

There isn't a single feature that is most valuable, but if I had to choose one, it would be the rich ability it gave us for making customized scripts. That was probably the most unique feature set for our situation. We used Control-M's Python Client and cloud data service integrations with AWS and, as a feature, it was very customizable. It gave us a lot of flexibility for customizing whatever data maneuver we wanted to do within a pipeline.

The Python Client and cloud data service integrations have a rich set of features with flexibility. It did not require additional, crazy skills or experience to deal with it. It was a nice transition into enabling a data scientist to leverage existing skills to build those pipelines.

Creating, integrating, and automating data pipelines with Control-M was straightforward. It did require some knowledge and training, but compared to other solutions, it was a lot simpler. Working with data workflows, with the data-coding language integrated into Control-M, was straightforward. The level of difficulty was somewhere between "medium" and "easy." It was not that hard to leverage existing skills and knowledge within this specific feature.

The user interface for creating, monitoring, and ensuring delivery of files as part of the data pipeline was very actionable. It was almost self-explanatory. Somebody with basic user-interface experience could navigate the calls to action and the configuration that is required. It was well-designed.

What needs improvement?

I would like to see them adopt more cloud. Most companies don't have a single cloud, meaning we have data sources that come from different cloud providers. That may have been solved already, but supporting Azure would be an improvement because companies tend not to have only AWS and GCP.

For how long have I used the solution?

I used it for a couple of years.

What do I think about the stability of the solution?

It's fairly stable. I don't recall any specific issues. 

What do I think about the scalability of the solution?

It's fairly scalable. For our needs, it scaled very nicely.

We have a shared model where we have a centralized, shared service organization when it comes to data. Different people will use it, but it's centralized.

How are customer service and support?

We used other solutions from BMC as well, and their customer support was always great. I give them a 10 out of 10.

Training or a Knowledge Base were available or you could ask a question by submitting a ticket.

How would you rate customer service and support?

Positive

Which solution did I use previously and why did I switch?

We had DataStage from IBM and SSIS.

The switch was really about streamlining the process. We had other tools that only did partial processes or were not doing it with the speed and efficiency that we were looking for. We were looking for a solution that could streamline things and solve 90 percent of our data challenges.

What was our ROI?

The analysis that I saw validated that the ROI was within a couple of years.

What's my experience with pricing, setup cost, and licensing?

The pricing was competitive, from what I understand.

Which other solutions did I evaluate?

We looked at continuing to use the same solutions we had been using, and there were a couple of other cloud-based solutions that we evaluated. One of them was Matillion. The ease of use was one component of our decision, as was the flexibility of scripting with Python. Those were the key differentiators.

What other advice do I have?

For the on-prem solution, we had to do the patching and whatever was required by the vendor, but the cloud implementation was a model that was usable. The upgrades, changes, and patching are done directly by the vendor.

Control-M was a critical piece of the puzzle, to help us with all the data transformation and projects that we had to do. It was part of either one specific project or even a larger project that required that middle data transformation so that we could get to analytics or any other consumption of that data.

Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Nagarajan Sankarammal - PeerSpot reviewer
Automation Architect at a tech vendor with 10,001+ employees
MSP
A highly capable, feature-rich solution with excellent third-party integrations
Pros and Cons
  • "The product has improved dramatically over the years; it offers a lot in terms of features and capabilities and integration with third-party tools. A wide range of models available with the product is critical in reducing manual and mundane work such as custom script writing. This saves significant amounts of time and, by association, money for the organization."
  • "Regarding product design and R&D, the DevOps pipeline could be improved with better capabilities and automation. API security and authentication is another area that could use improvement; users must have static credential passwords, which is a security concern."

What is our primary use case?

We use the solution to automate our business batches, workload processing, and some elements of our IT and system maintenance procedures and processes. These include sequential clips, programs, and workflows. We automate these and have them scheduled for regular execution. We needed an orchestrator, and Control-M fits our requirements well.

How has it helped my organization?

The availability of data and reports is vital, and the solution's capacity for timely processing and build generation improved considerably over time. As our operation grew, so did our use of Control-M, and there has never been a delay in the availability of data and reports, even with very high workloads. Eventually, we could also bring automated control over our back end. Control-M makes workflow orchestration simpler; it can deal with an impressive amount of transactions.

We realized the benefits of the solution a long time ago, and from time to time, there will be a situation that reminds us how valuable it is to us. Control-M is an overwhelmingly stable and steady product, free from issues and frequent disruptions. As is the case for any such tool, there are occasional bugs and fixes, but overall, it's a stable product and a fully integrated part of our operation.

What is most valuable?

The product has improved dramatically over the years; it offers a lot in terms of features and capabilities and integration with third-party tools. A wide range of models available with the product is critical in reducing manual and mundane work such as custom script writing. This saves significant amounts of time and, by association, money for the organization.

As a Control-M user for over 15 years, I see it as very straightforward to create, integrate, and automate data pipelines. Even in the beginning, when Control-M was more of a data architecture product, it was easy to pick up. I've seen multiple people adapting very well in terms of adapting and enabling the capabilities of the solution for business; it's straightforward. 

Ideally, agent lift modes of connectivity would be established on different platforms. We can get applications integrated directly with Control-M. That's a recent feature. There are ready-made platforms and plugins which allow us to see templates for workflow orchestration in third-party and custom in-house applications. It's a straightforward solution, and this is an area where Control-M excels.  

Our customers are pleased being with Control-M, despite some minor hiccups, which happen with any solution. They have been happy with the product for years, and it's an enterprise-wide batch workflow orchestration tool. That's how it is established in our organization and what our users are satisfied and familiar with. 

The process execution speed is excellent and has constantly improved over the years.

The bottom line is Control-M is a mission-critical solution, it's integral to our organization. 

What needs improvement?

Regarding product design and R&D, the DevOps pipeline could be improved with better capabilities and automation. API security and authentication is another area that could use improvement; users must have static credential passwords, which is a security concern.

The REST API supports FTP for file transfers, but we would like to see additional, more encrypted protocols and simplified file transfer encryption. Currently, the solution offers PGP encryption, which isn't the most straightforward. 

For how long have I used the solution?

We have been using the solution for around 15 years. 

What do I think about the stability of the solution?

The product is stable.

What do I think about the scalability of the solution?

The solution is scalable; we increased our usage over the years and plan to continue that. 

We have multiple teams at multiple geos and deployments; we're an enterprise-sized organization.

How are customer service and support?

Recently, there are some bugs with the product development, which necessitated R&D's involvement, which isn't ideal. We have fully integrated the solution into our production businesses, so any stability issues have a significant impact. There were cases where workarounds weren't provided quickly, with stubborn bugs needing environmental solutions.

How would you rate customer service and support?

Neutral

Which solution did I use previously and why did I switch?

I've worked with multiple other workload orchestration tools, including IMB Tivoli Workload Scheduler and a CA automation product. Control-M stands above the competitors in terms of stability. CA underwent an acquisition, leading to changes in product strategy and mergers with equivalent products like Automic, so Control-M was the surer option. It is also more robust and has greater system availability than the competitors.

How was the initial setup?

The initial deployment was complex, and this was by necessity. It is important to note that deployment is now more straightforward due to years of knowledge, experience, and newer features.

It took around two weeks to set up the Control-M infrastructure, and the process of bringing in business data and full adoption took place over years. It could be done faster, as in our case, there were other considerations involving budgeting, testing, and timelines. Setup of the initial infrastructure takes a few weeks, and then getting the platform running and configured can be done in a day or two. Further configuration and integration with LDAP and monitoring tools can take a little longer.

The solution can be managed and maintained by two or three staff members, but the number of staff involved in a deployment can vary significantly. It depends on the specific scenario and teams.

What was our ROI?

I would say we have a return on our investment; we have a vast amount of transactions and business automation implemented on a massive scale. We have adopted Control-M extensively, and it would be challenging to migrate to another solution in a reasonable time. We often look at alternatives but considering factors like timelines, resource availability, and team bandwidth, we keep coming back. The solution provides enormous value to our organization.

What's my experience with pricing, setup cost, and licensing?

The solution is not cheap, it comes with quite a hefty price tag. Control-M is the market leader, but we still want the price to be as friendly as possible. 

The solution comes with the base module and an additional one with a few extra plugins, which is helpful. 

Which other solutions did I evaluate?

We evaluate competitors yearly, but in terms of value for money, we always return to Control-M. We get an excellent return on our investment.

What other advice do I have?

I would rate this solution an eight out of ten.

I would advise any organization to do a proof of concept for their scenario before making a decision.

We don't currently use the Python Client, it's something we are planning to look into. We haven't started working on it, but we are in the review process to understand the client, and how it could fit into our operation.

The solution doesn't create new data as such, but it processes on top of the business data. 

We don't currently use the product for analytics, but we do plan to get Control-M data onto other systems for analytics and machine learning tasks.  

If we didn't have Control-M we would use an alternative solution. If there was a better one we would use that, or a product with a favorable cost and value proposition, which is a key factor.  

The tool always positively impacted our business, including our business service delivery speed. Over the years, there were one or two issues, but the vendor supporters could keep up. Some bugs required extensive development, and the support is excellent in this regard. They always have the right staff to assist during major productions or changes. Compared to before we had Control-M, it's as if we were previously traveling by foot, and now we've discovered the wheel. 

Regarding the audit preparation process, features like workload archiving come with an additional cost, which not all organizations can afford. I would instead maintain something locally on the system, but the solution is straightforward in terms of data necessities.

Control-M has to catch up in some areas, but it also offers specific capabilities and customization options. Application integration provides scope for exploration and deployment in custom developments. As a product supplier, BMC could focus on improving in areas indicated by their biggest customers. There is a lot of room for improvement.

File transfer support is Control-M's only significant limitation, as it only supports SSTP and STP transfers. Overall, other tools offer better security and file size in terms of file transfers. Therefore, the solution slows down when dealing with larger file sizes or a high volume of files, but it is sufficient for smaller organizations.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Buyer's Guide
Download our free Control-M Report and get advice and tips from experienced pros sharing their opinions.
Updated: July 2025
Buyer's Guide
Download our free Control-M Report and get advice and tips from experienced pros sharing their opinions.