Try our new research platform with insights from 80,000+ expert users
reviewer1376661 - PeerSpot reviewer
Sr. Data Engineer at a healthcare company with 10,001+ employees
Real User
Provides the ability to document primary/foreign key relationships and standardize them
Pros and Cons
  • "What has been useful, I have been able to reverse engineer our existing data models to document explicitly referential integrity relationships, primary/foreign keys in the model, and create ERDs that are subject area-based which our clients can use when working with our databases. The reality is that our databases are not explicitly documented in the DDL with primary/foreign key relationships. You can't look at the DDL and explicitly understand the primary/foreign key relationships that exist between our tables, so the referential integrity is not easily understood. erwin has allowed me to explicitly document that and create ERDs. This has made it easier for our clients to consume our databases for their own purposes."
  • "erwin generally fails to successfully reverse engineer our Oracle Databases into erwin data models. The way that they are engineered on our side, the syntax is correct from an Oracle perspective, but it seems to be very difficult for erwin to interpret. What I end up doing is using Oracle Data Modeler to reverse engineer into the Oracle data model, then forward engineer the DDL into an Oracle syntax, and importing that DDL into erwin in order to successfully bring in most of the information from our physical data models. That is a bit of a challenge."

What is our primary use case?

I am responsible for both a combination of documenting our existing data models and using erwin Data Modeler as a primary visual design tool to design and document data models that we implement for our production services.

My primary role is to document our databases using erwin to work with people and ensure that there is logically referential integrity from the perspective of the data models. I also generate the data definition language (DDL) changes necessary to maintain our data models and databases up to our client requirements in terms of their data, analytics, and whatever data manipulation that they want to do. I use erwin a lot.

It is either installed locally or accessed through a server, depending on where I have been. I have had either a single application license or pooled license that I would acquire when I open up erwin from a server.

How has it helped my organization?

We get data from many different sources where I work. We have many clients. The data is all conceptually related. There are primary subject area domains common across most of our clients. However, the physical sources of the data, or how the data is defined and organized, often vary significantly from client to client. Therefore, data modeling tools like erwin provide us with the ability to create a visual construct from a subject area perspective of the data. We then use that as a source to normalize the data conceptually and standardized concepts that are documented or defined differently across our sources. Once we get the data, we can then treat the data that has been managed somewhat disparately from a common conceptual framework, which is quite important.

At the moment, for what I'm doing, the interface to the physical database is really critical. erwin generally is good for databases. It is comfortable in generating a variety of versions of data models into DDL formats. That works fine.

What has been useful, I have been able to reverse engineer our existing data models to document explicitly referential integrity relationships, primary/foreign keys in the model, and create ERDs that are subject area-based which our clients can use when working with our databases. The reality is that our databases are not explicitly documented in the DDL with primary/foreign key relationships. You can't look at the DDL and explicitly understand the primary/foreign key relationships that exist between our tables, so the referential integrity is not easily understood. erwin has allowed me to explicitly document that and create ERDs. This has made it easier for our clients to consume our databases for their own purposes.

What is most valuable?

Its visualization is the most valuable feature. The ability to make global changes throughout the data model. Data models are reasonably large: They are hundreds, and in some cases thousands, of tables and attributes. With any data model, there are many attributes that are common from a naming perspective and a data type perspective. It is possible with erwin to make global changes across all of the tables, columns, or attributes, whether you are doing it logically or physically. Also, we use it to set naming standards, then attempt to enforce naming standards and changes in naming from between the logical version of the data models and the physical versions of the data models, which is very advantageous. It also provides the ability to document primary/foreign key relationships and standardize them along with being able to review conceptually the data model names and data types, then visualize that across fairly large data models.

The solution’s visual data models for helping to overcome data source complexity and enabling understanding and collaboration around maintenance and usage is very important because you can create or define document subject areas within enterprise data models. You can create smaller subsets to be able to document those visually, assess the integrity, and review the integrity of the data models with the primary clients or the users of the data. It can also be used to establish communications that are logically and conceptually correct from a business expert perspective along with maintaining the physical and logical integrity of the data from a data management perspective. 

What needs improvement?

We are not using erwin's ability to compare and synchronize data sources with data models in terms of accuracy and speed for keeping them in sync to the fullest extent. Part of it is related to the sources of the data and databases that we are now working with and the ability of erwin to interface with those database platforms. There are some issues right now. Historically, erwin worked relatively well with major relational databases, like Oracle, SQL Server, Informix, and Sybase. Now, we are migrating our platforms to the big data platforms: Hadoop, Hive, and HBase. It is only the more recent versions of erwin that have the ability to interface successfully with the big data platforms. One of the issues that we have right now is that we haven't been able to upgrade the version that we currently have of erwin, which doesn't do a very good job of interfacing with our Hive and Hadoop environments. I believe the 2020 version is more successful, but I haven't been able to test that. 

Much of what I do is documenting what we have. I am trying to document our primary data sources and databases in erwin so we have a common platform where we can visually discuss and make changes to the database. In the past couple of years, erwin has kind of supported importing or reverse engineering data models from Hive into erwin, but not necessarily exporting data models or forward generating the erwin-documented data models into Hive or Hadoop (based on my experience). I think the newest versions are better adapted to do that. It is an area of concern and a bit of frustration on my part at this time. I wish I had the latest version of erwin, either the 2020 R1 or R2 version, to see if I could be more successful in importing and exporting data models between erwin and Hive.

erwin generally fails to successfully reverse engineer our Oracle Databases into erwin data models. The way that they are engineered on our side, the syntax is correct from an Oracle perspective, but it seems to be very difficult for erwin to interpret. What I end up doing is using Oracle Data Modeler to reverse engineer into the Oracle data model, then forward engineer the DDL into an Oracle syntax, and importing that DDL into erwin in order to successfully bring in most of the information from our physical data models. That is a bit of a challenge. 

There are other characteristics of erwin, as far as interfacing directly with the databases, that we don't do. Historically, while erwin has existed, the problem is the people that I work with and who have done most of the data management and database creation are engineers. Very few of them have any understanding of data modeling tools and don't work conceptually from that perspective. They know how to write DDL syntax for whether it's SQL Server, Oracle, or Sybase, but they don't have much experience using a data modeling tool like erwin. They don't trust erwin nor would they trust any of its competitors. I trust erwin a lot more than our engineers do. The most that they trust the solution to do is to document and be able to see characteristics of the database, which are useful in terms of discussing the database from a conceptual perspective and with clients, rather than directly engineering the database via erwin. 

erwin is more of a tool to document what exists, what potentially will exist, and create code that engineers can then harvest and manage/manipulate to their satisfaction. They can then use it to make changes directly to our databases. Currently, when the primary focus is on Hive databases or Hadoop environment, where there is no direct engineering at this point between erwin and those databases, any direct or indirect engineering at the moment is still with our Oracle Database.

Buyer's Guide
erwin Data Modeler by Quest
December 2024
Learn what your peers think about erwin Data Modeler by Quest. Get advice and tips from experienced pros sharing their opinions. Updated: December 2024.
824,067 professionals have used our research since 2012.

For how long have I used the solution?

I have been using the solution on and off for 20 to 30 years.

What do I think about the stability of the solution?

It is pretty stable. Personally, I haven't run into any real glitches or problems with the output, the ability to import data when it does work correctly, the export/creation of DDL, or generation of reports.

We are trying to upgrade. This has been going on now for several months. We're trying to upgrade to the 2020 version. Originally, it was 2020 R1, but I think at this point people are talking about the 2020 R2 version. Now, I'm not part of our direct communications with erwin in regards to Data Modeler, but there are some issues that erwin is currently working on that are issues for my company. This have prevented us from upgrading immediately to the 2020 version.

What do I think about the scalability of the solution?

This gets down to how you do your data modeling. If you do your data modeling in a conceptually correct manner, scaling isn't an issue. If you don't do your data modeling very well, then you are creating unnecessary complexities. Things can get a bit awkward. This isn't an erwin issue, but more a consequence of who is using the product.

In the area that I'm working right now, I'm the only user. Within the company, there are other people and areas using the solution probably far more intimately in regards to their databases. I really don't know the number of licenses out there.

How are customer service and support?

The problem is that our issues are related to interfacing erwin Data Modeler with the Hadoop Hive environments. The issues have always been either what I was trying to do was not fully supported by our version of erwin Data Modeler. People have certainly tried to help, but there's only so much that they could tell me. So, it's been difficult. I am hoping that I can get back to people with some better answers once the newest version of erwin is available to us.

Which solution did I use previously and why did I switch?

The people who were previously responsible for the database development were very good engineers who knew how to write SQL. They could program anything themselves that they wanted to program. However, I really don't think that they really understood data modeling as such. They just wrote the code. Our code and models are still developing and not necessarily conformed to good data modeling practices. 

How was the initial setup?

In the past, I was involved in the initial setup. In traditional environments, it sets up pretty easily. In my current environment, where I'm trying to get it as intimately integrated with our big data platforms as possible, I'm finding it quite frustrating. However, I'm using an older version and think that is probably a significant part of the problem.

What was our ROI?

In other environments where I've worked, the solution’s ability to generate database code from a model for a wide array of data sources cuts development time. In this environment, erwin is not very tightly integrated into the development cycle. It is used more for documentation purposes at this point and for creating a nascent code which down the road gets potentially implemented. While it's not used that way at my current company, I think it would be better if it were, but there is a culture here that probably will prevent that from ever occurring.

What's my experience with pricing, setup cost, and licensing?

An issue right now would be that erwin doesn't have a freely available browser (that I am aware of) for people who are not data modelers or data engineers that a consumer could use to look at the data models and play with it. This would not be to make any changes, but just to visually look at what exists. There are other products out there which do have end user browsers available and allow them to access data models via the data modeling tool.

Which other solutions did I evaluate?

There is another tool now that people are using. It is not really a data modeling tool. It is more of a data model visualization tool, and that's SchemaSpy. We don't do data modeling with that. You get a visualization of the existing physical database. But that's where the engineers live, and that's what they think is great. This is a cultural, conceptual, understanding issue due to a lack of understanding and appreciation of what good data modeling tools do that I can't see changing based on the current corporate organization. 

What other advice do I have?

It is the only meaningful way to do any data modeling. It is impossible to conceptualize and document complex data environments and the integration between different data subject areas. You can write all the code or DDL you want, but it's absolutely impossible to maintain any sort of conceptual or logical integrity across a large complex enterprise environment without using a tool like erwin. 

You want to look at what you are trying to accomplish with erwin before implementing it.

  • Does the product have the ability to support or accomplish that?
  • Based on the technologies that you have decided you want to use to manage your data, how intimately does it integrate with those technologies? 

From my perspective of using the traditional relational databases, I think erwin probably works pretty well. 

For the newer database technologies, such as the Hadoop environment databases, it's not clear to me how successful erwin is. However, I'm not talking from the perspective of somebody who has been aggressively using the latest version. I don't have access to it, so I'm afraid my concerns or issues may not be valid at this point. I will find out when we finally implement the latest erwin version.

I would give the solution a seven or eight (out of 10).

Which deployment model are you using for this solution?

On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
IT Specialist at a government with 10,001+ employees
Real User
The fact that you can generate the DDL correctly from the model saves us a bunch of time
Pros and Cons
  • "The modeling portion of the tool is the most valuable. There are some notes, naming standards, and other functions that we use as well. There's a whole boatload of functionality in this thing and we use maybe 10% of it. It seems to be pretty common that not all the functionality is fully utilized. But it's just got gobs and gobs of stuff that you can implement if you so choose to."
  • "The only real complaint I have is the time it takes to do a database comparison on a large model. If they could speed that up, that would be the only thing I can think of that needs improvement."

What is our primary use case?

erwin is deployed on individual desktops and the individual users install it or have a help desk person install it for them.

Our primary use case is for during any type of project development or maintenance and application maintenance, we go through a process of modeling our data before it gets put into the database. We interact with the application development teams to determine what their requirements are and build the data models, and then turn them into actual physical database items.

How has it helped my organization?

erwin has definitely helped us improve our enforcement of standards and database design best practices. Before we really started using the tool or having a data modeling type of team, application development efforts all had their own database structures. Developers tend to not be too concerned with the data. They just want to make everything work for their application as easy as possible. Having the tool and having a team built around it has really helped us make sure that we're following the best normalization processes, we're not duplicating data, and we have a standard naming scheme that everybody has to follow.

What is most valuable?

The modeling portion of the tool is the most valuable. There are some notes, naming standards, and other functions that we use as well. There's a whole boatload of functionality in this thing and we use maybe 10% of it. It seems to be pretty common that not all the functionality is fully utilized. But it's got gobs and gobs of stuff that you can implement if you so choose to.

We've definitely expounded on the amount of features we use. They've built in some automated naming standards that have been really helpful for us. That's probably the biggest leap we've used. We've always used the comments and notes features, but the automated naming features have been very helpful.

Its ability to overcome data source complexity and enabling understanding and collaboration around maintenance and usage is extremely helpful because they give a visual to not only developers and database administrators, but the user base themselves. So the typical user isn't going to understand database functionality. Being able to show them a picture of how their data is actually going to look in the database is very helpful for their understanding of what we're trying to do with their data.

erwin's ability to compare and synchronize data sources with data models in terms of accuracy and speed for keeping them in sync is very good. We utilize that service quite a bit. The one drawback is if you have an extremely large complex model, the compare process can take quite a bit of time, more than four hours. 

Its ability to generate database code from a model for a wide array of data sources cuts development time. The fact that you can generate the DDL correctly from the model saves us a bunch of time. I would say it saves us around 40% to 50%. So even though you can generate the DDL, you still have to go in and tweak it a little bit. 

What needs improvement?

The only real complaint I have is the time it takes to do a database comparison on a large model. If they could speed that up, that would be the only thing I can think of that needs improvement.

For how long have I used the solution?

We've been using erwin since about 2000. We were using another product before, but it was way too cumbersome, so we switched to erwin.

What do I think about the stability of the solution?

Stability is excellent. It's been a solid product for years and I don't expect it to change.

What do I think about the scalability of the solution?

It's extremely scalable. Our environment has hundreds of tables. 

We have five data modelers using the tool. That's the team that actually works with the app dev and DBAs to actually come up with the database design. Then we have another five users that act more in a read-only type of mode. They just want to look at the data models, but they don't actually do any of the design work.

How are customer service and technical support?

Their support was excellent. Typically it has to do with going through the upgrade process. If we have an issue, we'll reach out to them. The other thing we've had to reach out to them about was the time it was taking to do a data comparison on our extremely large model to the actual physical database. They were very helpful and very professional.

We don't typically have problems transitioning between the models. We did last time, but it was actually an error on our end. It wasn't an error on the erwin end.

Which solution did I use previously and why did I switch?

We were using Cayenne. We switched because it was cumbersome.

How was the initial setup?

The initial setup is typically straightforward. Just follow their instructions and everything goes pretty smoothly.

For the Data Modeler portion itself, on each desktop, the setup took around half an hour, and we have around 10 desktops.

We didn't necessarily have a deployment strategy. We just gave the product to anybody that thought they needed it and let them run with it.

For maintenance, we need one person, but it's definitely not a full-time job. It's just adding and subtracting users and going through the upgrade process when we do that. As far as installation, everybody basically installs it themselves. We don't require a full-time person for that either.

We have a team around it, so if we add our data modeling team up, we use it about six hours a day per person. That would be about 18 hours a day for those guys. The read-only users rarely use it, so they're pretty insignificant. 

We probably only use 10% to 20% of the functionality and I don't see us expanding on that a whole lot. There's a lot of neat little things in there, but we don't have time to implement them all. There's some overhead that goes with those functions that we choose not to undertake.

Since we got a new guy on our team, he's gotten into some of those functions and has been able to utilize some of that stuff some more. We're actually probably closer to 30% or 40% of the functions at this point. We're not thinking about expanding because of the overhead. 

What's my experience with pricing, setup cost, and licensing?

I don't remember what our costs are. I know they just recently switched from a per seat type of licensing to a concurrent user type of licensing agreement, which is neither here nor there. I don't think it has increased or decreased the cost at all, but it's not obtrusive or invasive as far as the cost goes. It's fairly affordable.

There are also internal costs if you have hosted on-prem because you have to have a server and database to stand it up on.

Which other solutions did I evaluate?

We didn't evaluate another solution because I had used erwin at another location and was extremely familiar with it. And I had also used Visio and some more manual-type methods like Visio. At the point that we decided to switch over, I was confident that erwin was the best solution out there.

What other advice do I have?

erwin is by far the best tool I've ever used. 

My advice to somebody considering this solution is to go for it. It's easy. The functionality is fantastic. It's easy to pick up. It does basically everything you could want it to do.

The automation of reusable design rules and standards has helped us immensely once we implemented it because having the automated naming standards and things like that, we don't have to go in and think about it. We don't have to go in and physically type it. Between generating the DDL and getting it into physical implementation was saving us 40% to 50% of time. It's because of those automated features that that's happening as opposed to having to sit there and type out the DDL from scratch, it saves a ton of time.

It produces a time savings of about 40%.

The accuracy and the speed of this solution in transforming complex designs into well-aligned data sources absolutely make the cost of the tool worth it.

My advice would be to let things evolve over time. Start with the basics first. Just get into the ERD functions first and then start implementing some of the automated naming standards and things like that as you go. Otherwise, if you try to dive into the whole thing, you're just going to get overwhelmed because the product is so deep as far as features go. It's extremely intuitive. As far as the basics go, as far as getting your ERDs established, it's probably the easiest tool I've ever used. If you understand the basics of database design, it's extremely natural. If you have no clue about database design, then your learning curve is going to be large no matter what tool you pick. But erwin definitely cuts that learning curve down just because of its intuitiveness.

Once you start diving into the automated feature sets like naming standards and things like that, the learning curve there is a little steeper, but it's still not too bad. For a brand new person, if you try to delve into the automated stuff and all the additional functionality, you're just going to get overwhelmed and feel that there is too much overhead. But you don't need to implement all those features right off the bat.

I would rate erwin a ten out of ten.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
Buyer's Guide
erwin Data Modeler by Quest
December 2024
Learn what your peers think about erwin Data Modeler by Quest. Get advice and tips from experienced pros sharing their opinions. Updated: December 2024.
824,067 professionals have used our research since 2012.
reviewer1270548 - PeerSpot reviewer
VP Enterprise Data Architecture at a financial services firm with 5,001-10,000 employees
Real User
Top 20
Allows us to review databases with our business and technology people and to understand data relationships in our company
Pros and Cons
  • "The most valuable feature is the physical or visual representation of the database, showing the tables, the columns, the foreign keys, and the ability to generate DDL, so you can physically implement databases."
  • "I would like to see more support for working with the big-data world. There are so many new databases evolving and it's very hard for them to keep up with all of the new technologies. It would be good if they were able to dynamically support big-data platforms, other than Hive and Teradata."

What is our primary use case?

The whole purpose of the erwin tool is for the designing of databases. We use it for our conceptual, logical, and physical database modeling.

How has it helped my organization?

We've been using this product as long as I can remember at our company, so it's hard to say how it has improved things. It's existed since I've been here. But it gives everybody the ability to see the physical implementations in a visual manner.

The solution is extremely critical to driving business change and transformation in our company because we do 100 percent of our data modeling using this tool. We meet with the business to show what exists and we show them what our changes are going to be to meet new requirements. We review that with business to get its agreement to the approach. We also meet with technology to show how it's going to be transformed in the physical implementation. So it is extremely critical to our everyday process.

What is most valuable?

The most valuable feature is the physical or visual representation of the database, showing the tables, the columns, the foreign keys, and the ability to generate DDL, so you can physically implement databases.

It lets you display the actual physically implemented databases or the logical databases. That enables you to review them with business users or technology people, to understand the relationships of the data throughout the company and show how data is joined together to achieve whatever the desired business results are.

What needs improvement?

I would like to see more support for working with the big-data world. There are so many new databases evolving and it's very hard for them to keep up with all of the new technologies. It would be good if they were able to dynamically support big-data platforms, other than Hive and Teradata. There's a new release coming out this year and they're adding two more platforms in that next release. So they are striving to keep up with technology, but technology is just evolving too rapidly. There are just too many options.

For how long have I used the solution?

I've been using erwin since 1998 or 1999.

What do I think about the stability of the solution?

The solution is very stable. It continues to evolve. 

A lot of the things, a lot of the new tools that they're introducing as part of erwin, are to make it more of a data governance tool in general, beyond just the data modeling which we've traditionally used. That whole piece is rapidly evolving. I've been watching it evolve over the past two or three years. We're not ready to purchase the products yet that they're putting out because we feel things are still evolving, but in the next couple of years they'll be the leader in the entire data governance realm.

What do I think about the scalability of the solution?

Data Modeler is very scalable. It keeps evolving as new technologies come out. People put in requests for it to be able to support different database platforms, and in just about every release in the past couple of years they've come out with support for one or two additional platforms. They are trying to keep up with customers' demands. They're very good about continuing to upgrade support for their legacy stuff as well. They're evolving and they're doing a pretty good job.

How was the initial setup?

The product was sold by a company called Platinum when I first started using it. It was then sold to CA and I was involved when CA produced its first rollout. After being owned by CA, it then was spun off to its own, standalone company, as erwin. I was there for the initial deployment of that as well. So, I've done many deployments of erwin, as different releases have come out.

The setup has become more complex. That's probably related to the fact that they're doing more things on the cloud, such as licensing, which has caused problems because we have very tight security here. Access to servers outside of our firewall causes issues for people who work in certain regions around the world because we restrict access to the cloud for governance purposes. So, we have had some issues with licensing. People can't connect to the license server, because it's done over the cloud, so they have to do an off-line license, which locks the license. We have concurrent licenses. So when a license is locked, even though the person is no longer using the product, because that person is not connected through the cloud, erwin's system doesn't know to release the license. We've worked with erwin and they have tried to help mitigate that, but we still do encounter issues with licensing.

In terms of deployment, just the install of the product on somebody's machine takes about 10 minutes. It's not very long at all. There are other features, such as setting up users in Model Mart, which take longer because you have to analyze the user's needs and set up appropriate permissions. That could take longer, depending on what the user's roles are.

As for our implementation strategy for Data Modeler, we just deployed it on someone's computer. We tested it on that person's box, one that everybody had access to. We all got to try the tool to see that we wanted to use it and to understand its features. Once everybody was comfortable with the features of it, we then had to upgrade our Model Mart repository, which is where we store all of the erwin models. Everyone has to be on the same release. So, we have to QA the whole process of upgrading our Windows Server and upgrading our database server. After we do those upgrades, we can then deploy the software on the machines. 

That's probably one of the biggest issues: Everybody has to be on the exact same version and release to be able to work together, if you're using the Model Mart repository. It's not very backward-compatible.

What about the implementation team?

We did have to involve erwin consultants because of the firewall issues that we were having when we were doing testing. We had to involve their helpdesk. Their helpdesk is extremely responsive. They actually tried to help us immediately on the phone. We needed a higher level of support so they scheduled meetings where we were sharing screens with them and they were able to help us. They were very helpful. One of the best features of erwin is its helpdesk.

What was our ROI?

It's hard to know how to gauge ROI. We've been using it since I got here. With the tool, we have a very good service-oriented architecture. We know exactly where all the data is; it's very clearly documented. If we didn't have this tool, I don't know how we would manage knowing where data is or manage having a consistent business glossary or data dictionary.

Which other solutions did I evaluate?

We've experimented with other solutions, such as ER/Studio, which has had different names. We experimented with SAP PowerDesigner, but that was not as robust in performing what we wanted it to do.

The main differences between the products we evaluated and erwin would be the ease of use, between logical and physical transformed models. The logical is more on the business side, and the physical is more on the technical side. The ease of maintaining those two models together was the number-one advantage of erwin. Number-two is the ability of the tool to support many platforms while successfully creating DDL, without issues with the DDL. erwin also has the ability to do compares of the models against the database, and to synchronize differences, whether importing database changes into the model or exporting the model to the database. Those were the primary things that it did well.

But I come back to ease of use. It's a very easy tool to train somebody on and for them to use. ER/Studio is probably the second-best product, but it's not as self-explanatory, it's not as easy to use. It's a little bit more clunky. It probably performs just as well, but it's a bit more difficult to use.

What other advice do I have?

If you want good data architecture in your company, you need to have database design done. It's probably the most important factor for having things clearly modeled and documented. erwin Data Modeler is not just a modeling tool, it's also used for documentation. If you're using the tool's functions properly, analyzing the documentation, flagging fields that are NPPI data, it is invaluable for business use. You can generate data dictionaries, you can make sure people are speaking common languages, and you can enforce company standards so that people are doing things in a consistent manner. It's an invaluable tool. If you want to have good data architecture, you need to have a tool like this.

We don't currently use the collaborative web modeling capability. We just recently purchased that tool and we are planning on deploying it at the end of Q1 of this year.

We don't use the erwin data transformation for integration to a wider ecosystem. We are actually able to directly do all of the transformations that we need from erwin, so we're not required to do any transformations. It supports legacy systems like Db2, Oracle, SQL Server, and now Teradata and Hive, which were introduced in the past few years. But it can currently support all of the data modeling we need to support, so no transformations are needed.

We have different flavors of people who use the tool. We have people who are dedicated data architects, that's their full-time job. There are 15 to 20 of them in the company. And we have many people who do use it for very specific applications on more of a part-time basis, where they're doing the data modeling and reviewing it with an enterprise architect. There are about 150 people who are doing that. Overall, we have about 170 people who have access to the software.

For deployment, upgrades, and maintenance of the solution, we generally require four people. We require somebody to do a Windows upgrade; we require somebody to do a database upgrade, and that's for the Mart repository portion; and we have two people who do the testing for the erwin tool: somebody who installs the upgrades of erwin on the local machines, and somebody who's testing it. When it comes to the installs and the upgrades, each person who's using the tool is expected to do that on their own. We set up a deployment package and everyone runs it when they're told to execute the upgrade.

Which deployment model are you using for this solution?

On-premises
Disclosure: PeerSpot contacted the reviewer to collect the review and to validate authenticity. The reviewer was referred by the vendor, but the review is not subject to editing or approval by the vendor.
PeerSpot user
DominicMackenzie - PeerSpot reviewer
Senior data architect at a financial services firm with 51-200 employees
Real User
Top 5Leaderboard
The solution can be used for entity relationship data modeling, but it’s too rigid in terms of its theory
Pros and Cons
  • "Drag-and-drop data modeling and reverse engineering out of databases are the most valuable features of erwin Data Modeler by Quest."
  • "I would like the solution to be less rigid in terms of its theory."

What is our primary use case?

I use the solution for entity relationship data modeling to build the databases.

What is most valuable?

Drag-and-drop data modeling and reverse engineering out of databases are the most valuable features of erwin Data Modeler by Quest.

What needs improvement?

I would like the solution to be less rigid in terms of its theory. It supports the theory very well, but it's too rigid. It's focused on database design and theoretic database design. I'd like it to be able to do better extractions.

For how long have I used the solution?

I have worked with erwin Data Modeler by Quest for 20 years.

What do I think about the stability of the solution?

The solution is too rigid in terms of usability. The other thing is you can do things to make it crash, and when it crashes, the models created with it become read-only. To be honest, I prefer using Visual Paradigm.

I rate erwin Data Modeler by Quest a seven out of ten for stability.

How was the initial setup?

It took some coordination to get the license transferred to me. Otherwise, the solution's initial setup is straightforward.

What about the implementation team?

I've probably spent half a day not running the installation but waiting for emails back from the vendor.

What other advice do I have?

It's a good enough solution if the whole purpose of your data modeling is to generate databases. I also use data modeling for other purposes, and that's why I prefer Visual Paradigm. It allows me to do data modeling that's not so much focused on databases.

Overall, I rate erwin Data Modeler by Quest a seven out of ten.

Which deployment model are you using for this solution?

On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
Data Management & Automation Manager at a consultancy with 11-50 employees
Reseller
Saves a lot of development time
Pros and Cons
  • "The most valuable features are the ability to reverse engineer and do model comparison. With the reverse engineering, I can understand the databases from third-party products. With the model comparison, I can track the differences between two versions of the same database."
  • "I would like to have more data sources from other, different vendors. In recent years, the vendor has reduced the number of data sources, and I would like to have more data sources for every brand. For example, with Oracle, I would like to have compatibility for many versions, including old ones, not just the most recent."

What is our primary use case?

We usually use it to design new databases as well as reverse engineer some databases from third-party products, e.g., ERPs or monetary software.

What is most valuable?

The most valuable features are the ability to reverse engineer and do model comparison. With the reverse engineering, I can understand the databases from third-party products. With the model comparison, I can track the differences between two versions of the same database.

Because I can graphically see the Modeler database, that is very helpful for my job as it helps me understand the database. It is very different from SQL and DML scripts, which are very hard to understand with just sentences. When we have a graphic, that is very helpful. We can save time understanding that database.

I like the synchronization ability a lot because it can tell me to apply some level of governance to my models. I can be sure that the model in my documentation or development environment matches with the database that is working in our production environment. It is accurate. Though, it is not always fast when we have dozens of tables, but it works. I wait about an hour in order to have a big database synchronized.

The solution’s code generation ensures accurate engineering of data sources. It avoids rework.

What needs improvement?

I would like to have more data sources from other, different vendors. In recent years, the vendor has reduced the number of data sources, and I would like to have more data sources for every brand. For example, with Oracle, I would like to have compatibility for many versions, including old ones, not just the most recent. 

The technical support could be better. They could give faster solutions.

For how long have I used the solution?

I have been using this solution since 1995.

What do I think about the stability of the solution?

It is stable.

What do I think about the scalability of the solution?

Only when the database is too big, then we could have some trouble. We are talking about maybe 12,000 tables, then it starts to have some problems.

With erwin, we just need to add memory to the computer in order to work with bigger databases. However, it would be good to have erwin for other platforms, e.g., Linux and Macintosh, not just Windows. 

How are customer service and support?

The technical support is good. They are highly skilled. 

Which solution did I use previously and why did I switch?

Before erwin, I was manually using Notebook for my databases. Before erwin, I was designing databases and analyzing them manually all the time.

We chose erwin because it was the only solution which could help us design a database on the computer.

What was our ROI?

It saves a lot of development time. I think we are saving from two weeks to one month annually. It depends on the size and complexity of the database.

The solution’s automation of reusable design rules and standards is good compared to basic drawing tools. It saves time and keeps us from errors, which are very costly in the database. Therefore, we can get back our money very quickly.

The accuracy and speed of the solution in transforming complex designs into well-aligned data sources makes the cost of the tool worth it.

What's my experience with pricing, setup cost, and licensing?

erwin is expensive compared to other solutions. We are paying almost $6,000 per seat a month.

Which other solutions did I evaluate?

I have used different solutions along the way, but then I moved back to erwin. Besides erwin, I have tried IDERA Embarcadero, but I think erwin is more usable and has helped me to do my job better.

What other advice do I have?

I rate this solution as nine out of 10.

Which deployment model are you using for this solution?

On-premises
Disclosure: My company has a business relationship with this vendor other than being a customer: Reseller
PeerSpot user
reviewer2202039 - PeerSpot reviewer
Data Architect at a government with 10,001+ employees
Real User
Top 20
An easy-to-deploy data modeling solution that allows users to reuse entities and attributes
Pros and Cons
  • "The product allows us to reuse entities and attributes."
  • "We can only get licenses through partners."

What is our primary use case?

We use the solution for data modeling.

How has it helped my organization?

We believe that the solution would help provide some structure to our data modeling and data design practices.

What is most valuable?

The product allows us to reuse entities and attributes. Whether we're working on a conceptual or logical model, once we have defined them, we can reuse them rather than defining them all over again. It's a clever feature and helps with our work.

What needs improvement?

We can only get licenses through partners. We cannot purchase directly from Quest. The partners end up charging a big margin on top of the actual price even though they are not providing any service. It was a letdown.

For how long have I used the solution?

I have been using the solution for four months.

What do I think about the stability of the solution?

I rate the tool’s stability an eight out of ten. The tool has crashed a couple of times. The stability must be improved.

What do I think about the scalability of the solution?

I am the only person using the tool in my organization. We plan to increase the usage, but it is currently not being used much.

How are customer service and support?

The support personnel are a bit slow in responding. The vendor talks about providing 24/7 support, but it could take 24 to 36 hours to get a response. The communication is just via email. It slows down the whole troubleshooting process.

How would you rate customer service and support?

Neutral

How was the initial setup?

The initial setup was easy. I rate the ease of setup a ten out of ten. It took me five minutes to deploy the product. To deploy it, I downloaded the software, ran the executable, and it installed itself. I did the deployment myself.

What's my experience with pricing, setup cost, and licensing?

The solution is expensive. I rate the pricing a nine out of ten. The price is not fixed. The product does not have a clear pricing model.

Which other solutions did I evaluate?

We evaluated ER/Studio but did not like its pricing structure. They have different versions which require multiple licenses for the applications we use. It was not easy to use. We preferred the pricing model of erwin Data Modeler.

What other advice do I have?

The amount of data the tool works with is minimal, so scalability is irrelevant to the tool. It just uses metadata. People looking to use the solution must compare it with other tools like ER/Studio. ER/Studio and erwin Data Modeler are quite similar in the functionality they offer. It all comes down to what works for us in terms of pricing. Overall, I rate the solution an eight out of ten.

Which deployment model are you using for this solution?

On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
reviewer1754778 - PeerSpot reviewer
Lead Data Architect at a tech services company with 1,001-5,000 employees
Real User
Feature-rich, with the most recent integration technology
Pros and Cons
  • "They have a lot of features and the most up-to-date technology integration, which I haven't seen in other products."
  • "This is a very complex product."

What is our primary use case?

We used this solution for three to five projects that we had.

What is most valuable?

They have a lot of features and the most up-to-date technology integration, which I haven't seen in other products.

In terms of features, I believe they were doing very well in the latest technologies as well.

What needs improvement?

There are too many features. People find it extremely difficult to navigate. It's a great product, but there are too many features and no one knows exactly where things are. For example, the documentation is not always straightforward, which is why I use another competitor's product, which is simpler and easier to use. It has more features than the product I previously used.

This is a very complex product.

They have a lot of features, and that may be the reason why I'm not familiar with them and has made it more complicated for me. 

In comparison to other products I've used, such as ER studio, which is a competitor to erwin. It is simple and easy to code as well as other functions. Even the coding has become much simpler. I can see that this is a little more complicated. I'm not certain. Right now, I can explicitly state that no features are required because I still need to explore some areas. Maybe the reporting area will be interesting if it is more dynamic.

For how long have I used the solution?

I have been working with erwin Data Modeler (DM) for approximately four years.

We are using version 9 or 9.1.

What's my experience with pricing, setup cost, and licensing?

Pricing is very high compared to any other product.

What other advice do I have?

I would rate erwin Data Modeler (DM) a nine out of ten. 

Which deployment model are you using for this solution?

On-premises
Disclosure: I am a real user, and this review is based on my own experience and opinions.
PeerSpot user
Senior Consultant at a tech services company with 11-50 employees
Real User
Improves accuracy for generating target databases, allows us to pull metadata from a database, and makes it easy to display information and models
Pros and Cons
  • "Being able to point it to a database and then pull the metadata is a valuable feature. Another valuable feature is being able to rearrange the model so that we can display it to users. We are able to divide the information into subject areas, and we can divide the data landscape into smaller chunks, which makes it easier to understand. If you had 14 subject areas, 1,000 entities, and 6,000 columns, you can't quite understand it all at once. So, being able to have the same underlying model but only display portions of it at a time is extremely useful."
  • "I still use Visio for conceptual modeling, and that's mainly because it is easier to change things, and you can relax some of the rules. DM's eventual target is a database, which means you actually have to dot all the Is and cross all the Ts, but in a conceptual model, you don't often know what you're working with. So, that's probably a constraint with erwin. They have made it a lot easier, and they've done a lot, but there is probably still room for improvement in terms of the ease of presentation back to the business. I'm comparing it with something like Visio where you can change colors on a box, change the text color and that sort of stuff, and change the lines. Such things are a whole lot easier in Visio, but once you get a theme organized in erwin, you can apply that theme to all of the objects. So, it becomes easier, but you do have to set up that theme."

What is our primary use case?

In one of the companies, we used it as an information tool. We created a logical model so that the business would know what was in the offices down to the warehouse. The current use case is also the same. We have some places for information, so we can do a logical data model for them, but, usually, it would go towards building an actual database, which also involves reverse engineering of an existing one because people don't know what's in there.

It is currently on-prem, but we still have a separate server.

How has it helped my organization?

We want to bring different erwin components together and tell a business user story. So, having all of it on one platform to be able to tell one story makes it not as fragmented as components have been in the past. 

In my previous company, when we had 1,000 tables, 6,000 columns, and 14 subject areas, trying to explain to people in the organization was difficult. Without the tool, it would have been impossible. With the tool, it was a lot easier because you could show a steward how this is his or her domain. For each steward, you could say, "Well, this is your domain over here." Once they had that, they could understand what you were talking about. So, it improved communication. We had a point where two stewards were looking at the models, and one of them said, "I think that one that you've got over there is actually mine." The other one said, "I think you're right." So, we actually moved an entity from one subject area to another because now they had the ability to see what was in their subject area. They could go and see what wasn't theirs and should be someone else's. If we didn't have the tool, we wouldn't have that visibility and wouldn't have been able to recognize that sort of situation. 

Its ability to generate database code from a model for a wide array of data sources cuts development time. You don't have to re-key things. You put in the information at one spot, and it flows out from there. There are so many parameters you can put on the physical side. You can put in your indexes, and you can put in expected size changes. You can store all sorts of information within the model itself. It is a really good repository of all that sort of information, and then you just push a button, and it generates the other end. It works really well. In terms of time-saving, if you had to write it all out by hand, it would take weeks. It would probably take three or four times longer without the tool.

It certainly improves accuracy for the generation of target databases because you're only putting information in one spot. You don't have to retype it. For example, I saw the word conceptual model misspelled today. So, if you have to re-key something, no matter how careful you are, you're going to misspell things, which would cause problems down the track, whereas if you make a mistake in DM, there is only one place you have to go and fix it, and then, you would regenerate the downstream stuff. This means that you don't have to touch anything physical. You generate it, and then you can use it.

What is most valuable?

Being able to point it to a database and then pull the metadata is a valuable feature. Another valuable feature is being able to rearrange the model so that we can display it to users. We are able to divide the information into subject areas, and we can divide the data landscape into smaller chunks, which makes it easier to understand. If you had 14 subject areas, 1,000 entities, and 6,000 columns, you can't quite understand it all at once. So, being able to have the same underlying model but only display portions of it at a time is extremely useful.

I am currently trying to compare and synchronize data sources with data models, and it is pretty good. It shows you all the differences between the two systems. After that, it is a matter of what you want to do with them. It is certainly helpful for bringing models in and being able to compare. At the moment, I'm comparing something that's in a database with something that was in the DDL statement. So, these are two different sets of sources, and I can bring different sources together and compare them in the one, which is really helpful.

What needs improvement?

I still use Visio for conceptual modeling, and that's mainly because it is easier to change things, and you can relax some of the rules. DM's eventual target is a database, which means you actually have to dot all the Is and cross all the Ts, but in a conceptual model, you don't often know what you're working with. So, that's probably a constraint with erwin.

They have made it a lot easier, and they've done a lot, but there is probably still room for improvement in terms of the ease of presentation back to the business. I'm comparing it with something like Visio where you can change colors on a box, change the text color and that sort of stuff, and change the lines. Such things are a whole lot easier in Visio, but once you get a theme organized in erwin, you can apply that theme to all of the objects. So, it becomes easier, but you do have to set up that theme. I think they've got three to four initial themes. There is a default theme, and then there are two or three others that you can pick from. So, having more color themes would help. In Visio, you have a series of themes where someone who knows about color has actually matched the colors to each other. So, if you use the colors in the theme, they will complement each other. So, erwin should provide a couple more themes.

They could perhaps think of having an entry-level product that is priced a bit lower. For extra features, the users can pay more.

For how long have I used the solution?

I have been using it at least since 2003. I have used it at multiple organizations.

What do I think about the stability of the solution?

It has always been really stable in the different organizations that I've used it in. It has always been a pretty good product.

What do I think about the scalability of the solution?

It works fine with the number of people who have been using the product. We're talking about 10 to 12 people, not thousands of people. I haven't ever been in an organization where thousands of people even needed to get to the product. Probably the biggest drawback in scalability is the cost per seat rather than the actual product. The product works fine.

Our current organization has probably about 5 to 10 people using it. We're a consultancy, so we're using it in various roles. So, a lot of it is to do with understanding. As consultants, we try to understand what a client has in the organization and what sort of data they have to make sure there is actually data in the system that can answer their business questions. So, that's the sort of thing we use it for. We can turn around and give them designs. We can show what it is, and then we can turn around and make it what it would be. It is used by analysts and developers. They are not developing software. They are probably developing the database, but then, people would develop software.

I've used it on all the projects I've been on so far. I've been with this company for a short time, and it has come into play for pretty much all of the projects that I've been on. We want to use it more extensively. We want to use the erwin suite. We've got the modeler, but we also want to use their BI tool. We would like to evolve and come up with a story that links all of them together.

We have only just got the BI suite installed. We're starting to play around with it and see what we can do with it. We're doing some training on it at the moment. In a previous company also, somebody from erwin came to show it to us, and it was reasonably new at that point. That was last year. It is a reasonably new product. So, getting them to talk to each other has also been fairly new. erwin has only done it in the last couple of years. 

How are customer service and support?

I haven't had dealings with them, but the dealings I've had with erwin as a company have always been really good. So, I would rate them a nine or 10 out of 10.

Which solution did I use previously and why did I switch?

I use Visio on the conceptual side. We've got Informatica, and I think it has got a modeling component in there. We try to get a range of products because we're doing consulting in various organizations, and they have got various tools. Usually, it depends on what a client has already installed. Sometimes, it also depends on their budget. Something like Informatica is usually at the top right end corner of the Gartner Quadrant, but it could also be overkill for smaller organizations because the benefit may not be there. So, a lot of time, it is horses for courses. You have to sort of tailor any solution to meet a client's needs.

How was the initial setup?

I haven't ever really installed erwin. One of the other guys has done that. Most of the places had it installed already. Usually, the complexity depends on how the organization does its software deployment. So, you have to go and request the software and then somebody has to give you the package. Once you get the package, it is pretty straightforward. It is usually less of a problem on erwin's side and more of an issue with how an organization deploys any erwin software, but once you deploy it, it works fine.

Some places that I've worked with were very strict about doing testing on COTS products to make sure that there are no viruses on it and also to make sure that it plays nicely with the rest of the system. So, those sorts of organizations may take longer in terms of testing. You put it on a test machine first and make sure it is not going to kill anything. They might have to repackage some stuff before they put it out to the network. To deploy a vanilla thing, I would think that it would only take a couple of hours.

In terms of maintenance, at the moment, I think we've got one person. The main thing is deploying new versions. You've got a server stood up, and you have to put the software out there. I don't know if there is anything else beyond that.

What was our ROI?

We haven't done an ROI for the current version. When you look at the total cost of creating or understanding what you've currently got through reverse engineering, and you look at the total cost of creating new products and new databases and maintaining them over time, and then you put that into the return on investment model, it is well worth it.

The accuracy and speed of the solution in transforming complex designs into well-aligned data sources make the cost of the tool worth it. If you didn't have the tool and a single developer or a single modeler was trying to do the same thing, the speed would be three or four times slower. If you multiply that by the cost of that person and then you also consider the cost of the other people who are waiting for that person to create a database design, it multiplies out. So, it is well worth it.

What's my experience with pricing, setup cost, and licensing?

It has increased in price a fair amount over the years. It has always been expensive because it is a comprehensive product, and presumably, they have to do a tremendous amount of testing to make sure that everything works. It has always been dear because usually, a very specific target audience of data architects has the need for modelers, and not everyone in the organization would need to get a copy of it. Only people who are actually working in the database space need it. So, it has always been a very specialized piece of software, and it has been priced accordingly.

I don't specifically know what we're paying now. About three years ago, in another organization, I have this memory of 6,000 AUD a seat or something like that, but I am not sure. In the mid-2000s, it was something like 1,200 AUD a seat. I get the impression that there was a price jump when it was spun off from CA as a separate company, which is understandable, but it could sometimes be a barrier in some organizations picking it up.

I haven't talked to erwin people yet, but I'm going to suggest to them that they could perhaps think of having an entry-level product that is priced a bit lower, and then, you can buy the extra suite. That's what Microsoft does. They package a few things so that you have something, but if you want this extra stuff that has enterprise features, such as they talk to each other and have great bits and pieces, you have to pay more. I don't think there are any additional costs. It is per product, and there are different license levels. 

What other advice do I have?

Oracle Data Modeler, which is free, is one of the competitors that erwin has. You can't argue with the price point on that one, but erwin is much more comprehensive and easier to use. It is easier to display information and models to business people than something like Oracle Data Modeler, which does the job, but erwin does it a lot better. So, my advice would be that if you can afford it, get it.

Its visual data models have certainly improved over time in terms of overcoming data source complexity and enabling understanding and collaboration around maintenance and usage. It was originally designed as a tool to build databases with, and it retains a lot of that. It still looks like that in a lot of cases, but it has also been made more business-friendly with a sort of new front end. So, it used to be all or nothing where when you wanted to show somebody just the entity names or just the entity descriptions, you had to switch all of the entities on your diagram just to show names. Now, you can show some of them. You can shrink down some of them, and you can keep some of them expanded. So, it has become a more useful information-sharing tool over time. It is extremely helpful.

In my previous company, it was the enterprise data model, and you could paper a room with it if you printed the information out. To present that information to people, we had to chunk it down into subject areas. We had to present smaller amounts of information. Because it was linked to the underlying system, we could reuse the information that we had in a model in other models. The biggest lesson was to chunk the information down and present it in a digestible form rather than trying to show the entire thing because otherwise, people would run away screaming.

One of the places didn't have a modeling tool in it, and they were trying to do the documentation using Confluence. It was just a nightmare trying to keep it maintained with different developers using different tables and then needing to throw something into one and adding something into another one. It was just a nightmare. If they had one tool where they could put it all in one place, it would have been so much easier than the mess they had.

I would rate erwin Data Modeler a nine out of 10.

Which deployment model are you using for this solution?

On-premises
Disclosure: My company has a business relationship with this vendor other than being a customer: Partner
PeerSpot user
Buyer's Guide
Download our free erwin Data Modeler by Quest Report and get advice and tips from experienced pros sharing their opinions.
Updated: December 2024
Buyer's Guide
Download our free erwin Data Modeler by Quest Report and get advice and tips from experienced pros sharing their opinions.