Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
In Talend Administration Center environments with a high number of tasks (700–800) or frequent task cycles (e.g., executions every minute), the task execution history table in the Talend Administration Center (TAC) database may grow rapidly.
A large history table can negatively impact overall Talend Administration Center performance, leading administrators to manually truncate the table to maintain stability.
Talend Administration Center provides configuration parameters to automatically purge old execution history and maintain system performance.
TaskExecutionHistoryCleanUp
Talend Administration Center includes two configuration parameters in the Talend Administration Center database configuration table that control the cleanup process.
These parameters allow administrators to adjust:
Both parameters must be tuned to effectively control table size.
| Description | Interval (in seconds) between cleanup operations |
| Default value | 3600 (1 hour) |
| Behavior | Talend Administration Center performs a cleanup every 3600 seconds. Setting value to 0 disables automatic cleanup |
The default value is 3600 seconds, which corresponds to 1 hour. After the time defined in this parameter has elapsed, the system cleans up old task execution history and old misfired task execution records. This means the system performs the cleanup action in one hour, rather than immediately.
Lower the value to check more frequently the records that need to be deleted. Set 0 to disable the delete actions.
| Description | Maximum retention time (in seconds) before records are purged |
| Default value | 1296000 seconds (15 days), 15 days × 24 × 60 × 60 = 1,296,000 |
| Behavior | During each cleanup cycle, Talend Administation Center deletes records older than the retention period. |
Both parameters are documented below:
improving-task-execution-history-performances | Qlik Talend Help
It is finally here: The first public iteration of the Log Analysis app. Built with love by Customer First and Support.
"With great power comes great responsibility."
Before you get started, a few notes from the author(s):
Chapters:
01:23 - Log Collector
02:28 - Qlik Sense Services
04:17 - How to load data into the app
05:42 - Troubleshooting poor response times
08:03 - Repository Service Log Level
08:35 - Transactions sheet
12:44 - Troubleshooting Engine crashes
14:00 - Engine Log Level
14:47 - QIX Performance sheets
17:50 - General Log Investigation
20:28 - Where to download the app
20:58 - Q&A: Can you see a log message timeline?
21:38 - Q&A: Is this app supported?
21:51 - Q&A: What apps are there for Cloud?
22:25 - Q&A: Are logs collected from all nodes?
22:45 - Q&A: Where is the latest version?
23:12 - Q&A: Are there NPrinting templates?
23:40 - Q&A: Where to download Qlik Sense Desktop?
24:20 - Q&A: Are log from Archived folder collected?
25:53 - Q&A: User app activity logging?
26:07 - Q&A: How to lower log file size?
26:42 - Q&A: How does the QRS communicate?
28:14 - Q&A: Can this identify a problem chart?
28:52 - Q&A: Will this app be in-product?
29:28 - Q&A: Do you have to use Desktop?
Qlik Sense Enterprise on Windows (all modern versions post-Nov 2019)
*It is best used in an isolated environment or via Qlik Sense Desktop. It can be very RAM and CPU intensive.
The information in this article is provided as-is and to be used at own discretion. Depending on tool(s) used, customization(s), and/or other factors ongoing support on the solution below may not be provided by Qlik Support.
Optimizing Performance for Qlik Sense Enterprise - Qlik Community - 1858594
We often see new users implement Qlik Stitch in under five minutes. While the user experience for adding a data source and a destination is simple, there is a lot of complexity behind the scenes.
Let’s pull the curtains back and see how each record gets from its source to its destination through Qlik Stitch’s data pipeline.
The journey typically begins in a SaaS application or database, where it’s either pushed to Stitch via our API or through a webhook, or pulled on a schedule by the Singer-based replication engine that Qlik Stitch runs against data sources like APIs, databases, and flat files.
The image below outlines the internal architecture. For a full view, click on the image or download it.
The data’s next stop is the Import API, a Clojure web service that accepts JSON and Transit, either point-at-a-time or in large batches. The Import API does a validation check and an authentication check on the request’s API token before writing the data to a central Apache Kafka queue.
At this point, Qlik Stitch has accepted the data. Qlik's system is architected to meet our most important service-level target: don’t lose data. To meet this goal, we replicate our Kafka cluster across three different data centers and require each data point to be written to two of them before it is accepted. Should the write fail, the requester will try again until it’s successful.
Under normal conditions, data is read off the queue seconds later by the streamery, a multithreaded Clojure application that writes the data to files on S3 in batches, separated according to the database tables the data is destined for. We have the capacity to retain data in Kafka for multiple days to ensure nothing is lost in the event that downstream processing is delayed. The streamery cuts batches after reaching either a memory limit or an amount of time elapsed since the last batch. Its low-latency design aims to maximize throughput while guarding against data loss or data leaking between data sets.
Batches that have been written to S3 enter the spool, which is a queue of work waiting to be processed by one of our loaders. These Clojure applications read data from S3 and do any processing necessary (such as converting data into the appropriate data types and structure for the destination) before loading the data into the customer’s data warehouse. We currently have loaders for Redshift, Postgres, BigQuery, Snowflake, and S3. Each is a separate codebase and runtime because of the variation in preprocessing steps required for each destination. Operating them separately also allows each to scale and fail independently, which is important when one of the cloud-based destinations has downtime or undergoes maintenance.
All of this infrastructure allows us to process more than a billion records per day, and allows our customers to scale their data volumes up or down by more than 100X at any time. Qlik Stitch customers don’t need to worry about any of this, however. They just connect a source, connect a destination, and then let Qlik Stitch worry about making sure the data ends up where it needs to be.
For additional information, reference the official Stitch documentation on Getting Started.
Data connections to Qlik Cloud or Qlik Sense Enterprise on Windows can fail for several reasons. In this article we aim to provide you with a step-by-step guide on how to begin troubleshooting data connectivity errors, ruling out the most common root causes.
Content:
The first step in your investigation is to verify basic network connectivity to the database.
Work with your network administrator on both where necessary.
If the network and firewall settings are not the issue, verify if the database operates as expected.
Next, we verify if the credentials and settings in our connection are correct. The easiest way to do this is by logging in directly to the database using its native client.
Examples of on-premise database clients include:
For cloud-based databases:
The specific tools you use will depend on the cloud provider you are using and the type of database you have.
Verify if the following are correct in Qlik Sense (details may vary depending on your data connection):
Depending on your permissions, you can either edit/verify your connection directly in the Qlik Sense App Data Load Editor or in the Space.
Using the Data Load Editor:
Depending on your permissions, you can either edit/verify your connection directly in the Qlik Sense App Data Load Editor or in the Qlik Sense Management Console.
Using the Data Load Editor:
Using the Qlik Sense Management Console:
Once you have confirmed the connection details, you can test the connection in Qlik Sense.
Data connections generally fail during reload. Qlik Sense will log details of the error or preceding warnings in the corresponding reload log. These error and warning messages can then be used to look up solutions or troubleshoot further.
Search the warnings and errors in our knowledge base for known solutions. Begin your search here.
If you have completed all the above and have not found solutions in the Qlik knowledge base, reach out to the database vendor for additional support. They will be able to provide specific information on troubleshooting the issue with your specific database.
After upgrading to Qlik Sense Enterprise on Windows May 2023 (or later), the Qlik Sense Repository Service may cause CPU usage spikes on the central node. In addition, the central Engine node may show an increased average RAM consumption while a high volume of reloads is being executed.
The Qlik Sense System_Repository log file will read:
"API call to Engine service was successful: The retrieved 'StaticByteSize' value for app 'App-GUID' is: 'Size in bytes' bytes."
Default location of log: C:\ProgramData\Qlik\Sense\Log\Repository\Trace\[Server_Name]_System_Repository.txt
This activity is associated with the ability to see the Base memory size in the Qlik Sense Enterprise Management Console. See Show base memory size of apps in QMC.
The feature to see Base memory size can be disabled. This may require an upgrade and will require downtime as configuration files need to be changed.
Take a backup of your environment before proceeding.
If any issues occur, revert the changes by restoring the backed up files and open a ticket with Support providing the changed versions of repository.exe.config, capabilities.json mentioning this article.
Show base memory size of apps in QMC
QB-22795
QB-24301
Qlik Sense Enterprise on Windows May 2023, August 2023, November 2023, February 2024.
Long access times to Qlik Cloud or when performing reloads using the Direct Access Gateway may be caused by high network latency between your location and the AWS datacenter hosting the affected tenant.
A possible browser test to check the network connection from a computer to AWS can be found at AWS Latency Test.
This test does not concern Qlik Cloud's platform, and can therefore be used to determine if there are purely network-related issues.
This is just a first-step tool, not a conclusive one. A latency of 300ms or more might still not affect user experience in Qlik Cloud, but it's still worth using the tool quickly for a first assessment.
How to verify and measure transfer speed from Direct Access gateway to Qlik Cloud
Qlik Insight Advisor performs slowly. No errors are logged.
To optimize performance,
For additional performance or optimization questions, head over to the Qlik App Development forum, where your knowledgeable Qlik peers and our active Support Agents can better assist you.
This article covers how to improve performance when using the Qlik Replicate Source Endpoint DB2i in a task, specifically by enabling the AfterImageOnly Internal Parameter.
Enabling the AfterImageOnly option on a DB2 for i (DB2i) source endpoint in Qlik Replicate provides performance and efficiency benefits, particularly when running Apply Changes tasks.
Content
In DB2i journaling, every change to a record can be logged with:
By default, both images may be captured. The AfterImageOnly setting signals Qlik Replicate to capture only the after image of each change.
AfterImageOnly is an Internal Parameter. To set it:
No reload of the task or table is required. You can resume your task after saving the Endpoint.
AfterImageOnly does not fit every situation. To not use the parameter, if:
Environment
When Qlik Replicate runs on a newly provisioned Linux server, the system memory usage is abnormally high and reaches critical thresholds.
To resolve the issue, reduce or disable huge page allocation by updating the kernel parameter:
vm.nr_hugepages = 0
To make it persistent across reboots, update /etc/sysctl.conf. Then apply the change by "sudo sysctl -p"
The issue is related to the Linux kernel parameter vm.nr_hugepages, which defines the number of huge pages (typically 2 MB each) that are pre-allocated in memory.
In this case, the system was configured with:
vm.nr_hugepages = 24629
This reserves approximately 48 GB of memory (2M * 24629) for huge pages, significantly reducing the amount of available memory for other processes. If too many huge pages are reserved, the system may run low on regular memory pages, leading to performance degradation or instability.
00352619
Qlik Replicate provides the option to set different types of notification rules. These notifications then send emails when specific events occur during a task's runtime.
Every notification rule is set for a specific event of interest. You can define one or more notification rules for one or more tasks.
For detailed information on the different notification options, please refer to the section Notification settings in the Qlik Replicate Users guide
This article describes how to set up a notification rule that will send an email to a recipient list when a task faces latency:
To deactivate the notification, untick the active check box:
For better performance of Talend Runtime Server, you could uninstall Kar files to free up storage space.
This article briefly introuduces How to uninstall kar files from Talend Runtime Server
Prerequiste
For command
kar:uninstall
It is a command used to uninstall a KAR file (identified by a name).
By uninstall, it means that:
For instance, to uninstall the previously installed my-kar-1.0-SNAPSHOT.kar KAR file:
karaf@root()> kar:uninstall my-kar-1.0-SNAPSHOT
Talend Runtime Server
Please run the following karaf commands to clean Kar files from your Repository
#stop running artifact task
bundle:list |grep -i <artifact-name>
bundle:uninstall <artifact-name | bundle-id>
#clean task kar cache
kar:list |grep -i <artifact-name>
kar:uninstall <artifact-name>
For more information about KAR file, please refer to Apache Content
kar:* commands to manage KAR archives.
Some particular job is always failing with OutOfMemory error when using Cloud Engine.
Cloud Engine has a maximum memory usage of 8GB
If you are also looking for OutOfMemory Error Occurs when executing a job on Remote Engine, please have a look at this article as below:
OutOfMemory Error Occurs When Executing a job on Remote Engine
There are connection and Net Work Slow issues between Talend Cloud and Talend Remote Engine
The following errors are produced in the logs:
org.apache.http.conn.ConnectTimeoutException: Connect to pair.**.cloud.talend.com:443 [pair.**.cloud.talend.com/****,
Retry processing request to {s}-> pair.**.cloud.talend.com:443
org.apache.http.conn.ConnectTimeoutException: Connect to pair.**.cloud.talend.com:443 [pair.**.cloud.talend.com/****,
pair.**.cloud.talend.com/****, pair.**.
This service is used to communicate the status information for the Remote Engine such as its heartbeats and availability.
org.apache.http.conn.ConnectTimeoutException: Connect to pair.**.cloud.talend.com:443 [pair.**.cloud.talend.com/****,pair.**.cloud.talend.com/****, pair.**.cloud.talend.com/****] failed: Connect timed out
Retrying request to {s}->pair.**.cloud.talend.com:443
This service is used to communicate about task runs such as receiving, deploying, or undeploying events, and sending information about task run status and metrics.
Failed to connect to [https://msg.**.cloud.talend.com:443] after: 1 attempt(s) with Failed to perform GET on: https://msg.**.cloud.talend.com:443 as response was: Connect to **** [****] failed: Connection refused: connect, continuing to retry.
# In seconds
wait.for.connection.parameters.timeout=300
flow.deployment.timeout=300to 600# In seconds
wait.for.connection.parameters.timeout=600
flow.deployment.timeout=600
There are connection issues and Network slow between Talend Cloud and Talend Remote Engine.
If you access the Internet via a proxy, you need to add the following URLs to your allowlist.
Adding URLs to your proxy and firewall allowlist
There are JVM hanging errors in wrapper.log during job running for Talend Remote Engine Server
JVM appears hung: Timed out waiting for signal from JVM.
JVM did not exit on request, terminated
JVM exited in response to signal SIGKILL (9).
Unable to start a JVM
<-- Wrapper Stopped
When the CPU is high (time interval) the jvm exited and also you see JVM hanging errors in wrapper log.
This is because when the jvm is not getting response, the wrapper tries to restart the container and try to recover from the hung behavior.
This article outlines how Qlik NPrinting and Qlik Sense Enterprise on Windows consume resources.
Qlik NPrinting interacts with Qlik Sense when a metadata reload or a task execution is started (a report preview or an On-Demand request is essentially the same as a task execution).
For metadata reloads, Qlik NPrinting directly opens a connection with the Qlik Sense Server, and the following actions are performed on the Qlik Sense side:
The resource consumption is mainly on the Qlik Sense side. In particular, opening the app requires RAM, and removing selections/bookmarks/Session Access requires calculations that consume CPU.
When a task is run, Qlik NPrinting connects to the Qlik Sense Server, which performs the following:
This is quite expensive in terms of resources, especially if many filters are present (or Cycles/Levels/Pages) because a filter is essentially a selection on the app and each selection requires calculations and, therefore, CPU usage.
Moreover, Qlik NPrinting is multi-treading. This means that multiple instances of the same app can be opened in Qlik Sense at the same time. This enables Qlik NPrinting to execute the requests of filter applications and table exports in parallel. On the other side, having many applications opened at the same time represents a resource cost for Qlik Sense.
The maximum number of apps that can be opened at the same time equals the number of logical cores on the Qlik NPrinting Engine machine. You have to sum them up if more engines are installed.
The situation is more complex if Session Access is used. In this case, Qlik NPrinting cannot apply all the requests on the apps in the same session. The app must be closed and re-opened for each Qlik NPrinting user, with a high consumption of RAM and CPU on the Sense side. This also loses the possibility to apply many of the Qlik NPrinting optimizations for the report generation.
On the Qlik NPrinting side, some resources are used by the Engine service to keep in memory images, tables, and other values sent by Sense. NPrinting then needs to place these objects correctly on the template. Also, this action is performed by the Qlik NPrinting Engine service and it is more resource-consuming on the Qlik NPrinting side. Generally, we do not expect a problematic consumption in a supported NPrinting environment, unless very heavy and complex reports are generated or many tasks are run at the same time. The main resource request is on the Qlik Sense side.
Does Qlik NPrinting respect the Qlik Sense load balancing rules?
Qlik NPrinting uses the same proxy to connect as an end user. Load balancing rules will be respected.
Out of memory error is found in the Talend Management Center or Remote Engine Logs when executing a job on Remote Engine.
The job is consuming a lot of memory or the JVM Memory allocated to the job or remote engine is not enough to execute.
If you are looking for allocating more memory and OutOfMemory exception from Talend Studio side, please refer to these articles
Question
Why are there multiple IP addresses showing up when using the nslookup command for Talend Cloud URLs like "api.ap.cloud.talend.com"?
nslookup api.ap.cloud.talend.com
Multiple IP addresses show up because Round-Robin DNS is being used. It ensures high availability and distributes traffic among multiple servers.
What is Round-Robin DNS? Optimize Server Load
There are multiple ways to go about finding out the exact process of locking a file and preventing QlikView or Qlik Sense from carrying out a specific operation. This article covers one possible option on a Windows Server Operating Systems, Process Monitor.
For other options and third-party implementations, please contact your Windows administrator.
Common causes for locks are:
The following error is showing for tasks that were hanging in queue:
No engines were available to process this task. You can try to run the task manually now. Run this task during a time when your processors are available.
If you are running too many tasks in parallel, experiencing time out and tasks failed, please consider the following.
Too many tasks are running at the same time, and since the tasks that were in queue timed out, it failed to execute.
On a different note, please also review the steps to clean up the task logs on the Remote Engine
automatic-task-log-cleanup