Data Governance Senior Advisor at Abu Dhabi National Oil Company
Real User
Top 5
2024-06-19T13:03:00Z
Jun 19, 2024
We are implementing erwin to streamline data modeling and automate network changes for increased efficiency and comprehensive data development. This option simplifies the need for better data indexing.
Data Architect at a performing arts with 201-500 employees
Real User
Top 5
2023-11-15T17:30:05Z
Nov 15, 2023
I use the solution for reverse engineering as well as forward engineering. I do logical and physical modeling, and some of what I'm doing right now is reverse engineering from actual databases because they have no diagrams. The solution helps me diagram the current and help me design the future.
Learn what your peers think about erwin Data Modeler by Quest. Get advice and tips from experienced pros sharing their opinions. Updated: December 2024.
VP Enterprise Data Architecture at a financial services firm with 5,001-10,000 employees
Real User
Top 20
2023-02-21T17:07:00Z
Feb 21, 2023
Our primary use is for doing database designs on just about any platform. The main users are dedicated data architects, while we also have development team staff using the tools to review models. Additionally, our database admins access the solution for implementing the Data Definition Language (DDL).
Consultant at Beijing Essential Data Tech Co., Ltd
Reseller
Top 10
2023-02-03T13:16:27Z
Feb 3, 2023
Our company is the solution's only partner and reseller in China. We use the solution to provide data lineage to our customers' production environments. Most of our customers are in the mid-sized range.
Software Engineer Staff at Lockheed Martin Corporation
Real User
2021-12-31T20:33:42Z
Dec 31, 2021
I am using it for database design. I am using it to architect and generate one database platform from another. It involves reverse engineering and SQL generation.
Senior Consultant at a tech services company with 11-50 employees
Real User
2021-08-17T15:33:00Z
Aug 17, 2021
In one of the companies, we used it as an information tool. We created a logical model so that the business would know what was in the offices down to the warehouse. The current use case is also the same. We have some places for information, so we can do a logical data model for them, but, usually, it would go towards building an actual database, which also involves reverse engineering of an existing one because people don't know what's in there. It is currently on-prem, but we still have a separate server.
Senior Data Architect at a financial services firm with 10,001+ employees
Real User
2021-06-07T22:56:00Z
Jun 7, 2021
We use the Workgroup Edition for sharing data models across the organization. The primary reason we're using the Workgroup Edition over the Standard Edition is the centralized repository of models. Standard Edition requires individuals to determine how to share data models, whether you share them in a local LAN directory, through email, SharePoint, or Livelink. You have to come up with your own versioning scheme and your method for sharing models. With the Workgroup Edition, because it has the centralized repository with version control embedded in it, it standardizes how your organization does versioning and centralizes all the models at the same time.
Senior Data Warehouse Architect at a financial services firm with 1,001-5,000 employees
Real User
2020-12-29T10:56:00Z
Dec 29, 2020
We use erwin DM as a data modeling tool. All projects in the data warehouse area go through the erwin model first and get reviewed and get approved. That's part of the project life cycle. And then we exude the scripts out of DM into Snowflake, which is our target database. Any changes that happen after that also go through erwin and we then make a master copy of the erwin model. Our solution architecture for projects that involve erwin DM and Snowflake is an on-prem Data Modeler desktop version, and we have a SQL database behind it and that's where the models are stored. In terms of erwin Data Modeler, Snowflake is the only database we're using. We are not utilizing a complete round-trip from DM for Snowflake. We are only doing one side of it. We are not doing reverse-engineering. We only go from the data model to the physical layer.
Independent Consultant at a tech consulting company with 1-10 employees
Real User
2020-10-19T09:50:00Z
Oct 19, 2020
The use case was normally to update data model designs for transaction processing systems and data warehouse systems. Part of our group also was doing data deployment, though I personally didn't do it. The work I did was mostly for the online transaction systems and for external file designs. I didn't use it for data sources. I used the solution for generation of code for the target in the database. Therefore, I went from the model to the database by generating the DDL code out of erwin. We had it on-premise. There was a local database server on SQL, then we each had a client that we install on our machines.
Data Management & Automation Manager at a consultancy with 11-50 employees
Reseller
2020-10-19T09:50:00Z
Oct 19, 2020
We use it in order to create models, do some reverse engineering in the case of existing databases, and for comparing models, e.g., what is in the design vs reality.
My previous employer's use case was around data warehousing. We used it to house our models and data dictionaries. We didn't do anything with BPM, etc. The company that I left prior to coming to my current company had just bought erwin EDGE. Therefore, I was helping to see how we could leverage the integration between erwin Mapping Manager and erwin Data Modeler, so we could forward engineer our models and source port mappings, then mapping our data dictionary into our business definitions. We didn't use it to capture our sources. It was more target specific. We would just model and forward engineer our targets, then we used DM to manage source targets in Excel. Only when the company first got erwin EDGE did we start to look at leveraging erwin Mapping Manager to manage source targets, but that was still a POC. As far as early DM source specific, we didn't do anything with that. It was always targeted.
Data Modeler at a government with 10,001+ employees
Real User
2020-10-14T06:37:00Z
Oct 14, 2020
When I work from home, my use case for erwin is for when I get a request for a database upgrade. Usually, the request comes in with a whole bunch of tables and names so I'll go into the DM and I'll start building out what they're asking for. Once we actually get them to be able to view it and understand it, then we'll go back and forth with the developers and the requesters to make sure that it's exactly what they're looking for. We'll spend a few days making sure everything looks correct. Once that's finished, I'll send it out. Unfortunately, I can't do a PDF straight from erwin so I'll copy everything into Word and then save my Word as a PDF. With that PDF, I'll be able to send it off to all the stakeholders, not just the developers and the requesters, so that everybody can see it, even the ones that don't have erwin itself. My office use case is pretty much the same, except with the office, we add in Model Mart. We have our entire network, all the databases, and everything in Model Mart and it's over 1,500 different tables, relationships, attributes, and things like that. It's a really large model. Then, we break down that model into individual subject areas and we work through those. We go back to any new requests, we'll build them in Data Modeler and we'll go back and forth with the requesters, making sure everything looks like what they're expecting it to. They'll usually just send us either a spreadsheet of names and data types and then we build from there.
President at Global Retail Technology Advisors, LLC
Real User
Top 10
2020-07-26T08:19:00Z
Jul 26, 2020
I was part of a standards organization and we built a data model that is a standard data model for use in retail. That data model is now been released in version 7.3 and it is implemented all over the world. We don't implement the model, we've built the logical model and then the companies build their own physical model from there. erwin is a retail data model, which means that it handles the operational side of retail, which means there are somewhere around 8,000 attributes in it. It has got around 10 groupings of things. We have a grouping on transactions and there are all kinds of transactions that can occur in retail. The whole customer life cycle is covered in the inventory, items, and all that. The use case is for retail operations. It's massive. There are hundreds of use cases in this.
We have a couple of really important use cases for erwin. One of them is that we automate the pull of metadata from the repository itself, so that we have all the model metadata that we can then put into a centralized hub that we can access with other applications. Another reason we pull all the metadata out of the model is to run it through our model validation application, which telsl us if this model is healthy or not and if it meets our standards or not. The other use case that's really important is managing the abbreviations file that erwin uses to convert logical terms into physical terms. The way that you manage it today within erwin is very manual and you'll go from a spreadsheet, make changes, and upload, et cetera-- but we've created an API application where if we take the main standard file and keep it in the database, make the changes in the database, then we have an application that goes out into the Mart file, deletes the glossary, replaces it with the table from the database. It's all automated at the push of a button. It's things that would take us days to make changes in the standard files and do updates in eight different files.
Architecture Sr. Manager, Data Design & Metadata Mgmt at a insurance company with 10,001+ employees
Real User
2020-06-30T08:17:00Z
Jun 30, 2020
We use the erwin Data Modeler tool to document conceptual, logical, and physical data design. Business data models capture the understanding of the data from a business perspective, which can then drive physical design to ensure data is represented and used correctly.
Technical Consultant at a insurance company with 1,001-5,000 employees
Real User
2020-06-25T10:53:00Z
Jun 25, 2020
I'm an application developer with a fair amount of database background so I mostly use the tool to do physical modeling to support our application development. I'm a firm believer in not just adding columns to a table but to actually think about it, put together an Erwin model, and look at the relationships. I used to like to generate the model and generate changes all through the tool but being honest, one of my biggest frustrations with Erwin is that it's very difficult to forward engineer and keep things in sync. It used to be so easy and now it's very difficult. It's very frustrating to use this tool for that. We use it for data modeling but they also do a lot of logical modeling and architecture, and we also use the naming standards capability to force corporate standards across the models.
Sr. Data Engineer at a healthcare company with 10,001+ employees
Real User
2020-06-25T10:53:00Z
Jun 25, 2020
I am responsible for both a combination of documenting our existing data models and using erwin Data Modeler as a primary visual design tool to design and document data models that we implement for our production services. My primary role is to document our databases using erwin to work with people and ensure that there is logically referential integrity from the perspective of the data models. I also generate the data definition language (DDL) changes necessary to maintain our data models and databases up to our client requirements in terms of their data, analytics, and whatever data manipulation that they want to do. I use erwin a lot. It is either installed locally or accessed through a server, depending on where I have been. I have had either a single application license or pooled license that I would acquire when I open up erwin from a server.
Data Modeler at a logistics company with 10,001+ employees
Real User
2020-04-13T06:27:00Z
Apr 13, 2020
We use erwin to design conceptual, logical, and physical data models for new projects. We use a Forward Engineering tool to forward engineer data models into new database structures. We use the reverse engineering tool to bring databases into data models and erwin. We also generate HTML reports of the models to share with our customers. Whenever we do have a new project that requires a new approach, we do try using erwin for it. For example, if we have an XSD message file, then we would try to see if there is a way to get that into erwin for better visibility of the structures that we have to work with.
Technology Manager at a pharma/biotech company with 10,001+ employees
Real User
2020-03-22T06:49:00Z
Mar 22, 2020
The use cases are for our enterprise data warehouse where we have an enterprise model being maintained and we have about 11 business-capability models being maintained. Examples of business capabilities would be finance, human resources, supply-chain, sales and marketing, and procurement. We maintain business domain models in addition to the enterprise model. We're on-premise, a virtualized data center. We're running this as client-server, the client being PC-driven and the back-end for the erwin Mart is virtualized Windows Servers.
Enterprise Data Architect at a energy/utilities company with 1,001-5,000 employees
Real User
2020-02-13T07:51:00Z
Feb 13, 2020
We use it for our conceptual business-data model, for logical data modeling, and to generate physical database schemas. We also create dimensional modeling models.
EDW Architect/ Data Modeler at Royal Bank of Canada
Real User
2020-02-02T10:42:00Z
Feb 2, 2020
We work on different platforms like SQL Sever, Oracle, DB2, Teradata and NOSQL. When we take in requirements, it will be through Excel spreadsheet which is a Mapping Document and this contains information about Source and Target and there mapping and transformation rules. We understand the requirements and start building the conceptual model and then the logical model. When we have these Data Models built in erwin Data Modeler tool, we generate the PDF Data Model diagrams and take it to the team (DBA, BSAs, QA and others) to explain the model diagram. Once everything is reviewed, then we go on to discuss the physical Data Model. This is one aspect of the requirement from Data Warehouse perspective. Other aspect of the requirement can be from the operational systems where the application requirements might come through as DDLs with SQL extension files where we reverse engineer those files and have the models generated within erwin Data Modeler. Some of them, we follow the same templates as they are. But some others, once we reverse-engineer and have that Model within the erwin, we make changes to entity names, table names and capture metadata according to RBC standards. We have standards defined internally, and we follow and apply these standards on the Data Models.
Sr. Manager, Data Governance at a insurance company with 501-1,000 employees
Real User
2020-01-27T06:40:00Z
Jan 27, 2020
erwin Data Modeler does conceptual, logical, and physical database or data structure capture and design, and creates a library of such things. We use erwin Data Modeler to do all of the levels of analysis that a data architect does. We do conceptual data modeling, which is very high-level and doesn't have columns and tables. It's more concepts that the business described to us in words. We can then use the graphic interface to create boxes that contain descriptions of things and connect things together. It helps us to do a scope statement at the beginning of a project to corral what the area is that the data is going to be using. Then we do logical data models, which are completely platform-independent. They're only about datasets, the owned attribution, and different key analyses to determine what primary keys we want. And then we do database designs, which relates to the physical data models. We also do reverse-engineering where we are capturing the catalogs of existing systems, or purchased software, or even external vendor datasets. They send us data sets and we can reverse-engineer what they send us, especially the backup snapshots where a vendor in the cloud will send data as a backup restore. So to help the documentation for the reporting team, we do reverse-engineering so that they know what the table and column structure look like, along with sizing, nullability, and keys and constraints. erwin is on-prem. We have the Workgroup Edition, which means that we don't just have client-side software. We have client-side software that stores the data models back into a database which is on an on-prem server.
erwin pioneered data modeling, and erwin Data Modeler (erwin DM) remains trusted, award-winning software for data modeling and database design, automating complex and time-consuming tasks. Use it to discover and document any data from anywhere for consistency, clarity and artifact reuse across large-scale data integration, master data management, metadata management, Big Data, business intelligence and analytics initiatives – all while supporting data governance and intelligence efforts.
We are implementing erwin to streamline data modeling and automate network changes for increased efficiency and comprehensive data development. This option simplifies the need for better data indexing.
I use the solution for reverse engineering as well as forward engineering. I do logical and physical modeling, and some of what I'm doing right now is reverse engineering from actual databases because they have no diagrams. The solution helps me diagram the current and help me design the future.
We use the solution for data modeling.
I use the solution for entity relationship data modeling to build the databases.
I use erwin Data Modeler for a metadata project. I don't have thousands of tables to manage or a data warehouse or anything.
We use the product for data modeling.
Our primary use is for doing database designs on just about any platform. The main users are dedicated data architects, while we also have development team staff using the tools to review models. Additionally, our database admins access the solution for implementing the Data Definition Language (DDL).
Our company is the solution's only partner and reseller in China. We use the solution to provide data lineage to our customers' production environments. Most of our customers are in the mid-sized range.
We used this solution to upload and document our database models from the legacy systems.
We used this solution for three to five projects that we had.
I am using it for database design. I am using it to architect and generate one database platform from another. It involves reverse engineering and SQL generation.
We usually use it to design new databases as well as reverse engineer some databases from third-party products, e.g., ERPs or monetary software.
In one of the companies, we used it as an information tool. We created a logical model so that the business would know what was in the offices down to the warehouse. The current use case is also the same. We have some places for information, so we can do a logical data model for them, but, usually, it would go towards building an actual database, which also involves reverse engineering of an existing one because people don't know what's in there. It is currently on-prem, but we still have a separate server.
We use the Workgroup Edition for sharing data models across the organization. The primary reason we're using the Workgroup Edition over the Standard Edition is the centralized repository of models. Standard Edition requires individuals to determine how to share data models, whether you share them in a local LAN directory, through email, SharePoint, or Livelink. You have to come up with your own versioning scheme and your method for sharing models. With the Workgroup Edition, because it has the centralized repository with version control embedded in it, it standardizes how your organization does versioning and centralizes all the models at the same time.
We use erwin DM as a data modeling tool. All projects in the data warehouse area go through the erwin model first and get reviewed and get approved. That's part of the project life cycle. And then we exude the scripts out of DM into Snowflake, which is our target database. Any changes that happen after that also go through erwin and we then make a master copy of the erwin model. Our solution architecture for projects that involve erwin DM and Snowflake is an on-prem Data Modeler desktop version, and we have a SQL database behind it and that's where the models are stored. In terms of erwin Data Modeler, Snowflake is the only database we're using. We are not utilizing a complete round-trip from DM for Snowflake. We are only doing one side of it. We are not doing reverse-engineering. We only go from the data model to the physical layer.
The use case was normally to update data model designs for transaction processing systems and data warehouse systems. Part of our group also was doing data deployment, though I personally didn't do it. The work I did was mostly for the online transaction systems and for external file designs. I didn't use it for data sources. I used the solution for generation of code for the target in the database. Therefore, I went from the model to the database by generating the DDL code out of erwin. We had it on-premise. There was a local database server on SQL, then we each had a client that we install on our machines.
We use it in order to create models, do some reverse engineering in the case of existing databases, and for comparing models, e.g., what is in the design vs reality.
My previous employer's use case was around data warehousing. We used it to house our models and data dictionaries. We didn't do anything with BPM, etc. The company that I left prior to coming to my current company had just bought erwin EDGE. Therefore, I was helping to see how we could leverage the integration between erwin Mapping Manager and erwin Data Modeler, so we could forward engineer our models and source port mappings, then mapping our data dictionary into our business definitions. We didn't use it to capture our sources. It was more target specific. We would just model and forward engineer our targets, then we used DM to manage source targets in Excel. Only when the company first got erwin EDGE did we start to look at leveraging erwin Mapping Manager to manage source targets, but that was still a POC. As far as early DM source specific, we didn't do anything with that. It was always targeted.
When I work from home, my use case for erwin is for when I get a request for a database upgrade. Usually, the request comes in with a whole bunch of tables and names so I'll go into the DM and I'll start building out what they're asking for. Once we actually get them to be able to view it and understand it, then we'll go back and forth with the developers and the requesters to make sure that it's exactly what they're looking for. We'll spend a few days making sure everything looks correct. Once that's finished, I'll send it out. Unfortunately, I can't do a PDF straight from erwin so I'll copy everything into Word and then save my Word as a PDF. With that PDF, I'll be able to send it off to all the stakeholders, not just the developers and the requesters, so that everybody can see it, even the ones that don't have erwin itself. My office use case is pretty much the same, except with the office, we add in Model Mart. We have our entire network, all the databases, and everything in Model Mart and it's over 1,500 different tables, relationships, attributes, and things like that. It's a really large model. Then, we break down that model into individual subject areas and we work through those. We go back to any new requests, we'll build them in Data Modeler and we'll go back and forth with the requesters, making sure everything looks like what they're expecting it to. They'll usually just send us either a spreadsheet of names and data types and then we build from there.
I was part of a standards organization and we built a data model that is a standard data model for use in retail. That data model is now been released in version 7.3 and it is implemented all over the world. We don't implement the model, we've built the logical model and then the companies build their own physical model from there. erwin is a retail data model, which means that it handles the operational side of retail, which means there are somewhere around 8,000 attributes in it. It has got around 10 groupings of things. We have a grouping on transactions and there are all kinds of transactions that can occur in retail. The whole customer life cycle is covered in the inventory, items, and all that. The use case is for retail operations. It's massive. There are hundreds of use cases in this.
We have a couple of really important use cases for erwin. One of them is that we automate the pull of metadata from the repository itself, so that we have all the model metadata that we can then put into a centralized hub that we can access with other applications. Another reason we pull all the metadata out of the model is to run it through our model validation application, which telsl us if this model is healthy or not and if it meets our standards or not. The other use case that's really important is managing the abbreviations file that erwin uses to convert logical terms into physical terms. The way that you manage it today within erwin is very manual and you'll go from a spreadsheet, make changes, and upload, et cetera-- but we've created an API application where if we take the main standard file and keep it in the database, make the changes in the database, then we have an application that goes out into the Mart file, deletes the glossary, replaces it with the table from the database. It's all automated at the push of a button. It's things that would take us days to make changes in the standard files and do updates in eight different files.
We use the erwin Data Modeler tool to document conceptual, logical, and physical data design. Business data models capture the understanding of the data from a business perspective, which can then drive physical design to ensure data is represented and used correctly.
I'm an application developer with a fair amount of database background so I mostly use the tool to do physical modeling to support our application development. I'm a firm believer in not just adding columns to a table but to actually think about it, put together an Erwin model, and look at the relationships. I used to like to generate the model and generate changes all through the tool but being honest, one of my biggest frustrations with Erwin is that it's very difficult to forward engineer and keep things in sync. It used to be so easy and now it's very difficult. It's very frustrating to use this tool for that. We use it for data modeling but they also do a lot of logical modeling and architecture, and we also use the naming standards capability to force corporate standards across the models.
I am responsible for both a combination of documenting our existing data models and using erwin Data Modeler as a primary visual design tool to design and document data models that we implement for our production services. My primary role is to document our databases using erwin to work with people and ensure that there is logically referential integrity from the perspective of the data models. I also generate the data definition language (DDL) changes necessary to maintain our data models and databases up to our client requirements in terms of their data, analytics, and whatever data manipulation that they want to do. I use erwin a lot. It is either installed locally or accessed through a server, depending on where I have been. I have had either a single application license or pooled license that I would acquire when I open up erwin from a server.
We use erwin to design conceptual, logical, and physical data models for new projects. We use a Forward Engineering tool to forward engineer data models into new database structures. We use the reverse engineering tool to bring databases into data models and erwin. We also generate HTML reports of the models to share with our customers. Whenever we do have a new project that requires a new approach, we do try using erwin for it. For example, if we have an XSD message file, then we would try to see if there is a way to get that into erwin for better visibility of the structures that we have to work with.
The use cases are for our enterprise data warehouse where we have an enterprise model being maintained and we have about 11 business-capability models being maintained. Examples of business capabilities would be finance, human resources, supply-chain, sales and marketing, and procurement. We maintain business domain models in addition to the enterprise model. We're on-premise, a virtualized data center. We're running this as client-server, the client being PC-driven and the back-end for the erwin Mart is virtualized Windows Servers.
We use it for our conceptual business-data model, for logical data modeling, and to generate physical database schemas. We also create dimensional modeling models.
We work on different platforms like SQL Sever, Oracle, DB2, Teradata and NOSQL. When we take in requirements, it will be through Excel spreadsheet which is a Mapping Document and this contains information about Source and Target and there mapping and transformation rules. We understand the requirements and start building the conceptual model and then the logical model. When we have these Data Models built in erwin Data Modeler tool, we generate the PDF Data Model diagrams and take it to the team (DBA, BSAs, QA and others) to explain the model diagram. Once everything is reviewed, then we go on to discuss the physical Data Model. This is one aspect of the requirement from Data Warehouse perspective. Other aspect of the requirement can be from the operational systems where the application requirements might come through as DDLs with SQL extension files where we reverse engineer those files and have the models generated within erwin Data Modeler. Some of them, we follow the same templates as they are. But some others, once we reverse-engineer and have that Model within the erwin, we make changes to entity names, table names and capture metadata according to RBC standards. We have standards defined internally, and we follow and apply these standards on the Data Models.
erwin Data Modeler does conceptual, logical, and physical database or data structure capture and design, and creates a library of such things. We use erwin Data Modeler to do all of the levels of analysis that a data architect does. We do conceptual data modeling, which is very high-level and doesn't have columns and tables. It's more concepts that the business described to us in words. We can then use the graphic interface to create boxes that contain descriptions of things and connect things together. It helps us to do a scope statement at the beginning of a project to corral what the area is that the data is going to be using. Then we do logical data models, which are completely platform-independent. They're only about datasets, the owned attribution, and different key analyses to determine what primary keys we want. And then we do database designs, which relates to the physical data models. We also do reverse-engineering where we are capturing the catalogs of existing systems, or purchased software, or even external vendor datasets. They send us data sets and we can reverse-engineer what they send us, especially the backup snapshots where a vendor in the cloud will send data as a backup restore. So to help the documentation for the reporting team, we do reverse-engineering so that they know what the table and column structure look like, along with sizing, nullability, and keys and constraints. erwin is on-prem. We have the Workgroup Edition, which means that we don't just have client-side software. We have client-side software that stores the data models back into a database which is on an on-prem server.
The whole purpose of the erwin tool is for the designing of databases. We use it for our conceptual, logical, and physical database modeling.