We performed a comparison between OpenText Silk Test and OpenText UFT Developer based on real PeerSpot user reviews.
Find out in this report how the two Functional Testing Tools solutions compare in terms of features, pricing, service and support, easy of deployment, and ROI."A good automation tool that supports SAP functional testing."
"The scalability of the solution is quite good. You can easily expand the product if you need to."
"The ability to develop scripts in Visual Studio, Visual Studio integration, is the most valuable feature."
"The feature I like most is the ease of reporting."
"The major thing it has helped with is to reduce the workload on testing activities."
"Scripting is the most valuable. We are able to record and then go in and modify the script that it creates. It has a lot of generative scripts."
"The statistics that are available are very good."
"The most valuable feature is the Object Model, where you can directly pull up the object as a global or a local."
"There are many good things. Like it is intuitive and scripting was easy. Plus the availability of experienced resources in India due to its market leadership."
"This tool is really good. We don't need to write any code, but it writes the code itself, only record and play. And it is simple, and it is not heavy; I mean, it doesn't have a large footprint, and it works well for us."
"The solution helps to accelerate software testing automation. It will help to reduce lead time and increase productivity and efficiency."
"The most valuable features are the object repository."
"The most valuable feature is the automation of test cases."
"The most valuable feature for me is the number of protocols that can be tested. It not only tests Web, but also SAP, Siebel, .Net, and even pdf."
"The solution is very scalable."
"The solution has a lack of compatibility with newer technologies."
"The pricing is an issue, the program is very expensive. That is something that can improve."
"They should extend some of the functions that are a bit clunky and improve the integration."
"The support for automation with iOS applications can be better."
"Could be more user-friendly on the installation and configuration side."
"Everything is very manual. It's up to us to find out exactly what the issues are."
"We moved to Ranorex because the solution did not easily scale, and we could not find good and short term third-party help. We needed to have a bigger pool of third-party contractors that we could draw on for specific implementations. Silk didn't have that, and we found what we needed for Ranorex here in the Houston area. It would be good if there is more community support. I don't know if Silk runs a user conference once a year and how they set up partners. We need to be able to talk to somebody more than just on the phone. It really comes right down to that. The generated automated script was highly dependent upon screen position and other keys that were not as robust as we wanted. We found the automated script generated by Ranorex and the other key information about a specific data point to be more robust. It handled the transition better when we moved from computer to computer and from one size of the application to the other size. When we restarted Silk, we typically had to recalibrate screen elements within the script. Ranorex also has some of these same issues, but when we restart, it typically is faster, which is important."
"With Smart Bear products generally, you can have only one instance of the tool running on a machine."
"The parallel execution of the tests needs improvement. When we are running tests in LeanFT, there are some limitations in terms of running the same tests simultaneously across different browsers. If I'm running a test, let's say to log in, I should be able to execute it through IE, through Microsoft Edge, through Chrome, through Mozilla, etc. This capability doesn't exist in LeanFT. Parallel execution of the test cases across different browsers need to be added."
"Easier connectivity and integration with SAP would be helpful."
"It is unstable, expensive, inflexible, and has poor support."
"The pricing could be improved."
"I have to keep the remote machine open while the tests are running, otherwise, it leads to instability."
"In the next release, I would like to see the connectivity improved to be less complex and more stable."
"The support for .NET Framework and Visual Studio in Micro Focus UFT Developer is currently limited. At present, only Visual Studio 2019 is supported, despite the release of a newer version (2022). Similarly, the tool only supports up to .NET Framework version 4.3.8, while there have been six newer versions released. This is an area that could be improved upon, particularly in the Windows environment."
Earn 20 points
OpenText Silk Test is ranked 26th in Functional Testing Tools while OpenText UFT Developer is ranked 16th in Functional Testing Tools with 34 reviews. OpenText Silk Test is rated 7.6, while OpenText UFT Developer is rated 7.4. The top reviewer of OpenText Silk Test writes "Stable, with good statistics and detailed reporting available". On the other hand, the top reviewer of OpenText UFT Developer writes "Integrates well, has LeanFT library, and good object detection ". OpenText Silk Test is most compared with Selenium HQ, OpenText UFT One, Apache JMeter, froglogic Squish and OpenText LoadRunner Professional, whereas OpenText UFT Developer is most compared with OpenText UFT One, Tricentis Tosca, Original Software TestDrive, froglogic Squish and Selenium HQ. See our OpenText Silk Test vs. OpenText UFT Developer report.
See our list of best Functional Testing Tools vendors, best Test Automation Tools vendors, and best Regression Testing Tools vendors.
We monitor all Functional Testing Tools reviews to prevent fraudulent reviews and keep review quality high. We do not post reviews by company employees or direct competitors. We validate each review for authenticity via cross-reference with LinkedIn, and personal follow-up with the reviewer when necessary.