Head of Data Practice at a tech services company with 51-200 employees
Jan 23, 2023
I like modeling and code generation. It has become a pretty handy tool because of its short ideation to delivery time. From the time you decide you are modeling a data warehouse, and once you finish the modeling, it generates all the code, generates all the tables. All you have to do is tick a few things, and you can produce a fully functional warehouse. I also like that they have added all the features I have asked for over four years.
There were many valuable features, such as extracting any data to put in the cloud. For example, Qlik was able to gather data from SAP and extract SAP data from the platforms.
One of the most valuable features of this tool is its automation capabilities, allowing us to design the warehouse in an automated manner. Additionally, we can generate Data Lifecycle Policies (DLP) reports and efficiently implement updates and best practices based on proven design patterns.
The solution has room for improvement in the ETL. They have an ETL, but when it comes to the monitoring portion, Qlik Compose doesn't provide a feature for monitoring.
Head of Data Practice at a tech services company with 51-200 employees
Jan 23, 2023
It would be better if the first level of technical support were a bit more technically knowledgeable to solve the problem. I think they could also improve the injection of custom scripts. It is pretty difficult to add additional scripts. If the modeling doesn't give you what you want, and you want to change the script generated by the modeling, it is a bit more challenging than in most other products. It is very good with standard form type systems, but if you get a more complicated data paradigm, it tends to struggle with transforming that into a model.
Director - Metrics & Analytics at a computer software company with 1,001-5,000 employees
Nov 3, 2023
It could enhance its capabilities in the realm of self-service options as currently, it is more suited for individuals with technical proficiency who can create pages using it.
When processing data from certain tables with a large volume of data, we encounter significant delays. For instance, when dealing with around one million records, it typically takes three to four hours. To address this, I aim to implement performance improvements across all tables, ensuring swift processing similar to those that are currently complete within seconds. The performance issue primarily arises when we analyze the inserts and updates from the source, subsequently dropping the table. While new insertions are handled promptly, updates are processed slowly, leading to performance issues. Despite consulting our Qlik vendors, they were unable to pinpoint the exact cause of this occurrence. Consequently, I am seeking ways to optimize performance within Qlik Compose, specifically concerning updates.
Consultant at a tech services company with 1,001-5,000 employees
May 19, 2022
For more complex work, we are not using Qlik Compose because it cannot handle very high volumes at the moment. It needs the same batching capabilities that other ETL tools have. We can't batch the data into small chunks when transforming large amounts of data. It tries to do everything in one shot and that's where it fails.