Hello,
I have two stage processing pipeline:
1. Spark streaming job receives data from kafka and saves  it to
partitioned orc
2. There is spark etl job that  runs ones per day that compact each
partition( i have two variables for partitioning
dt=20180529/location=mumbai ( merge small files  to bigger one). Argument
for compactor job is full path to partition, so compactor job can not
update metadata.
So next time I want to read this table as orc ( if i try to read it as a
hive table it works ), spark read metadata directory,  found a structure of
orc table ( partitions and files that are placed into these partitions)  ,
tries to read some  file and fails with file not found, because  compactor
job had already removed this file and merged it to another file. I see
three workarounds
1. Remove _spark_metadata manually
2. modify spark compactor job the way when it updates metadata
3. found a configuration property that turns on ignoring of spark metadata
1 and 2 are good, but  it can be that I do not have access rights
So does the 3 chose exist ( I checked this
https://github.com/apache/spark/blob/56e9e97073cf1896e301371b3941c9307e42ff77/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/FileStreamSource.scala#L199
and could not find any property) ? If its not  I think it should be added
in some way to spark. May be it should not be a global property but a
property for query.

Reply via email to