Periodically Clean Up File Folders Using SSIS Productivity Pack
Applications and processes often rely on external resources, which can be expensive to retrieve or obtain. Therefore, modern applications often employ cache strategies to save these external resources in the local file system and achieve improved performance. In some other cases, you might have an application that keeps creating random temporary files to facilitate certain system requirements which cannot be achieved otherwise. However, as every coin has two sides, these caches or temporary files, in the long run, can build up in the system and occupy large disk space, which could have adverse effects and result in volume deficiency in the system. Implementing a scheduled cleanup of the cache folder by removing those files that are older than a certain age can become necessary in your ETL process.
Using Premium File Watcher Task to Monitor File Changes in an ETL Process
In the general integration practice, it is a very common use case to monitor a folder path for any file changes, and as soon as a file with certain criteria lands in the folder or gets modified, a process should kick off or otherwise, an email notification should be sent via email. With the new Premium File Watcher Task added to our SSIS Productivity Pack in our most recent v23.1 release, this can be easily achievable in an automated fashion. In this blog post, we will demonstrate the design and configurations required to watch a file path on a continuous basis and send an email notification when a specific change happens in it.
Working with Embedded XML Content in an ETL Process
In some unique cases, we may receive an XML document that needs to be processed in the ETL process, which might contain general XML content but includes some embedded XML content made of escaped characters. The integration requires the embedded XML content to be fully parsed and processed. In this blog post, we will show you how to properly handle and transform mixed XML content with escaped characters to ensure the successful parsing of API responses and files.
Implementing Retry Logic in Custom HTTP Requests Using KingswaySoft
There are many reasons you might need to send retries on failed HTTP-based service requests, particularly when you are working with API integration in an ETL process. Now, a simple retry can easily be done in the HTTP Connection Manager using its built-in Retry on Intermittent Errors option. In addition, there might be cases in which you would want to implement a more customized retry on additional HTTP errors to facilitate a more resilient integration process. In this blog post, we will see how KingswaySoft can help in both scenarios.
Split and Parse Through JSON Structures using KingswaySoft
Having a JSON response that needs to be properly parsed is a very common use case. This can be an advanced JSON structure like arrays, key-value pairs, etc. Our KingswaySoft JSON components can easily parse through them using the various settings and options available. However, there could be cases in which a JSON design is such that it could have records that need to be handled separately. In this blog post, we will discuss such a case.
Using KingswaySoft to Send Emails with Attachments Dynamically
Sometimes we need to attach files with specific conditions to one email and send it. In this blog post, we will show how you can quickly do it by using our components.
Extracting CRM Entity Metadata using Premium OData Components
In this blog, we are trying to get the metadata information from Dynamics 365 CRM API and save the entities and attributes in different tables. This is useful in many use cases where you wish to get a list of all the present entities and their corresponding metadata as a part of your integration - especially for further parameterization plans.
Unzip and Read Files from Cloud Storage for Transparent ETL Processing
In cases where you have a set of files compressed as a zipped folder in a cloud location, you might have requirements where you need to unzip these files first and then read them one by one to parse the data out to a destination database table or another set of files. To perform this manually wouldn't be as simple as it sounds. However, by using solutions provided by us, this can be easily automated.