Hey @27207436a3fb2f15a430, the default backing dataset for batch ontology objects from Pipeline Builder should have parquet files. Can you share what you see when you go to your object → open → view backing dataset and then in the dataset view, go to the details tab → files?
See screenshots below for the steps outlined above (you’ll see in the second screenshot there is a list of .parquet files)
The issue is that the materialization dataset is not giving you incremental updates which means you would first need a downstream incremental transform to detect deltas.
What’s your latency requirements for the end2end data flow? I have seen lightweight pipeline builder been able to keep op with <2 minutes latency requirements. That would mean you don’t require streaming datasets.
What’s consuming your data at the end of the pipeline?