Hi Experts, I have been using dataset.scanner to read the data with specific filter conditions and batch_size of 1000 to read the data.
ds.scanner(filter=pc.field('a') != 3, batch_size=1000).to_batches() I would like to know if it is possible to skip the specific set of batches, for example, the first 10 batches and read from the 11th Batch. https://arrow.apache.org/docs/python/generated/pyarrow.dataset.Dataset.html#pyarrow.dataset.Dataset.scanner Also, what's the fragment_scan_options in dataset scanner and how do we make use of it? Really appreciate any input. thanks! Regards, Alex