Incremental refresh created duplicates

I have an SQL source table with records that are never deleted, but are added and updated. There is a primary key column id.

In QuickSight, I added this as a simple Spice dataset, with an incremental refresh based on updated_at with a one hour window, triggered hourly.

But the dataset keeps adding “new” records, i.e. I have ~70K records instead of ~55K in the source. If I run a full refresh, the correct number of records is shown.

How can I tell QuickSight to respect the primary key and only update the records in an incremental refresh?

Hi @Gabriel_R ,

The incremental refresh is based on the lookback window , following video explains what happens : 2022 QuickSight Learning Series: Keep your data fresh with incremental SPICE refresh - YouTube

Kind regards,
Koushik

1 Like

Thanks for the video Koushik, make sense. I’m horrified :unamused:

So it’s by design that records are duplicated?!

PS: See at 18m00s in the video where Karthik explains that duplicates can occur.

This makes incremental refresh useless for so many use cases.

Please consider making this simpler and more powerful, i.e. similar to how Zoho Analytics does it:

Screenshot 2023-05-31 at 18.30.26 Zoho Analytics Importing data into an existing table from cloud - YouTube

Hi @Gabriel_R ,

The records are not duplicated by design , it’s looking into lookback window configuration. Hence the update is not similar to what happens in a datawarehouse scenario .

Thanks for your feedback, I will tag this question as a feature request so that we capture your requirements.

Kind regards,
Koushik