We now know how we can select the correct data, which type of tables we should use with lookups and how we can ensure that we only read through relevant datasets.
In practice it is still often the case that you must select a large and/or non-defined amount of data from the database, which should then be aggregated in accordance with specific rules for the high-performance reading.
Table of Contents
Different approaches are taken depending on the aggregation process. When adding, calculating the tally or the average, it is usually the case that you have to trawl through each dataset in the internal tables. It can also be quicker, however, when calculating the minimum or maximum value. You have to first sort the datasets contained before then cleaning with the instruction DELETE ADJACENT DUPLICATES. Using the additional COMPARING function, you also have the possibility to determine which fields should be taken into consideration when carrying out the comparison.
This is emphasised using the following example. We have a dbTab data source. This includes, among other things, 3 key fields, a lookup field for reading and a date field. It is our goal to only select the lookup value for the last, most current dataset, so that we can subsequently perform a more efficient reading.
We first select all the above mentioned fields from the dbTab, for which a key combination is available in our result_package.
SELECT
key_field1
key_field2
key_field3
date_field
lookup_field
FROM dbTab INTO TABLE iTab
FOR ALL ENTRIES IN result_package
WHERE key_field1 = result_package-key_field1
AND key_field2 = result_package-key_field2 AND key_field3 = result_package-key_field3
This means that we have several date fields for each key field, from which we only require the most current.
In this way we have ensured that the internal table iTab looks like this.
Following the instruction DELETE ADJACENT DUPLICATES, only the relevant datasets are left over and we can look forward to a high-performance reading of the datasets. :)
DELETE ADJACENT DUPLICATES FROM iTab COMPARING key_field1 key_field2 key_field3 date_field.
Who is b.telligent?
Do you want to replace the IoT core with a multi-cloud solution and utilise the benefits of other IoT services from Azure or Amazon Web Services? Then get in touch with us and we will support you in the implementation with our expertise and the b.telligent partner network.
Exasol is a leading manufacturer of analytical database systems. Its core product is a high-performance, in-memory, parallel processing software specifically designed for the rapid analysis of data. It normally processes SQL statements sequentially in an SQL script. But how can you execute several statements simultaneously? Using the simple script contained in this blog post, we show you how.
Many companies with SAP source systems are familiar with this challenge: They want to integrate their data into an Azure data lake in order to process them there with data from other source systems and applications for reporting and advanced analytics. The new SAP notice on use of the SAP ODP framework has also raised questions among b.telligent's customers. This blog post presents three good approaches to data integration (into Microsoft's Azure cloud) which we recommend at b.telligent and which are supported by SAP.
First of all, let us summarize the customers' requirements. In most cases, enterprises want to integrate their SAP data into a data lake in order to process them further in big-data scenarios and for advanced analytics (usually also in combination with data from other source systems).
As part of their current modernization and digitization initiatives, many companies are deciding to move their data warehouse (DWH) or data platform to the cloud. This article discusses from a technical/organizational perspective which aspects areof particularly important for this and which strategies help to minimize anyrisks. Migration should not be seen as a purely technical exercise. "Soft" factors and business use-cases have a much higher impact.