Dear colleagues,
I am reaching out for your assistance in configuring data filtering in the BigTales system. I have been grappling with this issue for the past two days and feel like I am at my wit’s end.
Task Description:
I need to establish dependent data filtering based on the following criteria:
- By operation type
- By counterparty
- By client
Requirements:
- Dependent Filtering: Selecting one parameter should affect the available options for the subsequent filters.
- Summation Post-Filtering: After completing the filtering process, I need to extract the sum of the filtered data.
Performance Issues with Large Datasets:
While both approaches work well for smaller datasets (up to 10,000 rows), they encounter significant performance issues as the dataset size increases:
- Up to 10,000 rows: Works perfectly fine.
- Around 500,000 rows: Starts showing noticeable performance degradation.
- 7 million rows: The methods fail completely due to excessive computational demands.
Specific Problems Encountered:
- Join List Limitations:
- I am unable to determine if there are limitations on the number of tokens or entries that can be stored in a Joined List. As the dataset grows, it becomes increasingly difficult to manage large lists efficiently.
- Split Text Performance:
- The Split Text function also starts to perform poorly with larger datasets, leading to delays and potential timeouts.
- Calculated Columns:
- Calculated columns begin to hang or become unresponsive as the number of rows increases. This significantly impacts the usability and reliability of the filtering process.