These are coming from the parquet library and as far as I know can be safely ignored.
On Mon, Sep 22, 2014 at 3:27 AM, Andrew Ash <and...@andrewash.com> wrote: > Hi All, > > I'm seeing the below WARNINGs in stdout using Spark SQL in Spark 1.1.0 -- > is this warning a known issue? I don't see any open Jira tickets for it. > > Sep 22, 2014 10:13:54 AM INFO: parquet.hadoop.InternalParquetRecordReader: > at row 0. reading next block > Sep 22, 2014 10:13:54 AM INFO: parquet.hadoop.InternalParquetRecordReader: > block read in memory in 6 ms. row count = 453875 > Sep 22, 2014 10:13:55 AM WARNING: parquet.hadoop.ParquetRecordReader: Can > not initialize counter due to context is not a instance of > TaskInputOutputContext, but is > org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl > Sep 22, 2014 10:13:55 AM INFO: parquet.hadoop.InternalParquetRecordReader: > RecordReader initialized will read a total of 454101 records. > Sep 22, 2014 10:13:55 AM INFO: parquet.hadoop.InternalParquetRecordReader: > at row 0. reading next block > Sep 22, 2014 10:13:55 AM INFO: parquet.hadoop.InternalParquetRecordReader: > block read in memory in 6 ms. row count = 454101 > Sep 22, 2014 10:13:55 AM WARNING: parquet.hadoop.ParquetRecordReader: Can > not initialize counter due to context is not a instance of > TaskInputOutputContext, but is > org.apache.hadoop.mapreduce.task.TaskAttemptContextImpl > > > Thanks! > Andrew >