Well anyways, even from Hue if I try loading partition wise data, it throws
the same error. Am really really perplexed to what this bug really is..
Although, if I try viewing the data in general, it displays me column names
/ values  / analysis etc but not partitionwise.

Thanks,
Sumit

On Sat, Aug 6, 2016 at 10:33 AM, Sumit Khanna <sumit.kha...@askme.in> wrote:

> Hey,
>
> I am having a parquet dir and a table mounted on it. the table is showing
> sample view , via hue fine but a simple query like select * from tablename
> gives this error :
>
>
>    - Bad status for request TFetchResultsReq(fetchType=0, operationHandle=
>    TOperationHandle(hasResultSet=True, modifiedRowCount=None,
>    operationType=0, operationId=THandleIdentifier(
>    secret='\xf7\xe7\x90\x0e\x85\x91E{\x99\xd1\xdf>v\xf7\x8c`',
>    guid='\xcc\xd6$^\xac{M\xaf\x9c{\xc2\xcf\xf3\xc6\xe7/')),
>    orientation=4, maxRows=100): TFetchResultsResp(status=TStatus(errorCode=0,
>    errorMessage='java.io.IOException: parquet.io.ParquetDecodingException:
>    Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_
>    mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-
>    r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet',
>    sqlState=None, infoMessages=['*org.apache.hive.service.cli.
>    HiveSQLException:java.io.IOException: parquet.io.ParquetDecodingException:
>    Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_
>    mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-
>    r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:25:24',
>    'org.apache.hive.service.cli.operation.SQLOperation:
>    getNextRowSet:SQLOperation.java:352', 'org.apache.hive.service.cli.
>    
> operation.OperationManager:getOperationNextRowSet:OperationManager.java:220',
>    'org.apache.hive.service.cli.session.HiveSessionImpl:
>    fetchResults:HiveSessionImpl.java:685', 'sun.reflect.
>    GeneratedMethodAccessor63:invoke::-1', 'sun.reflect.
>    DelegatingMethodAccessorImpl:invoke:DelegatingMethodAccessorImpl.java:43',
>    'java.lang.reflect.Method:invoke:Method.java:498',
>    'org.apache.hive.service.cli.session.HiveSessionProxy:
>    invoke:HiveSessionProxy.java:78', 'org.apache.hive.service.cli.
>    session.HiveSessionProxy:access$000:HiveSessionProxy.java:36',
>    
> 'org.apache.hive.service.cli.session.HiveSessionProxy$1:run:HiveSessionProxy.java:63',
>    'java.security.AccessController:doPrivileged:AccessController.java:-2',
>    'javax.security.auth.Subject:doAs:Subject.java:422',
>    'org.apache.hadoop.security.UserGroupInformation:doAs:
>    UserGroupInformation.java:1657', 'org.apache.hive.service.cli.
>    session.HiveSessionProxy:invoke:HiveSessionProxy.java:59',
>    'com.sun.proxy.$Proxy22:fetchResults::-1',
>    'org.apache.hive.service.cli.CLIService:fetchResults:CLIService.java:454',
>    'org.apache.hive.service.cli.thrift.ThriftCLIService:
>    FetchResults:ThriftCLIService.java:672', 'org.apache.hive.service.cli.
>    thrift.TCLIService$Processor$FetchResults:getResult:TCLIService.java:1553',
>    'org.apache.hive.service.cli.thrift.TCLIService$Processor$
>    FetchResults:getResult:TCLIService.java:1538', 'org.apache.thrift.
>    ProcessFunction:process:ProcessFunction.java:39', 'org.apache.thrift.
>    TBaseProcessor:process:TBaseProcessor.java:39',
>    'org.apache.hive.service.auth.TSetIpAddressProcessor:process:
>    TSetIpAddressProcessor.java:56', 'org.apache.thrift.server.
>    TThreadPoolServer$WorkerProcess:run:TThreadPoolServer.java:285',
>    
> 'java.util.concurrent.ThreadPoolExecutor:runWorker:ThreadPoolExecutor.java:1142',
>    
> 'java.util.concurrent.ThreadPoolExecutor$Worker:run:ThreadPoolExecutor.java:617',
>    'java.lang.Thread:run:Thread.java:745', 
> '*java.io.IOException:parquet.io.ParquetDecodingException:
>    Can not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_
>    mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-
>    r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:29:4',
>    
> 'org.apache.hadoop.hive.ql.exec.FetchOperator:getNextRow:FetchOperator.java:507',
>    
> 'org.apache.hadoop.hive.ql.exec.FetchOperator:pushRow:FetchOperator.java:414',
>    'org.apache.hadoop.hive.ql.exec.FetchTask:fetch:FetchTask.java:140',
>    'org.apache.hadoop.hive.ql.Driver:getResults:Driver.java:1670',
>    'org.apache.hive.service.cli.operation.SQLOperation:
>    getNextRowSet:SQLOperation.java:347', 
> '*parquet.io.ParquetDecodingException:Can
>    not read value at 0 in block -1 in file hdfs://askmehadoop/parquet1_
>    mpdm_mpdm_store/partitioned_on_seller_mailer_flag=1/part-
>    r-00000-a77c308f-c088-4f41-ab07-0c8e0557dbe1.gz.parquet:36:7',
>    'parquet.hadoop.InternalParquetRecordReader:nextKeyValue:
>    InternalParquetRecordReader.java:228', 'parquet.hadoop.
>    ParquetRecordReader:nextKeyValue:ParquetRecordReader.java:201', '
>    org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper:<
>    init>:ParquetRecordReaderWrapper.java:122', '
>    org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper:<
>    init>:ParquetRecordReaderWrapper.java:85', '
>    org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat:
>    getRecordReader:MapredParquetInputFormat.java:72',
>    'org.apache.hadoop.hive.ql.exec.FetchOperator$FetchInputFormatSplit:
>    getRecordReader:FetchOperator.java:673', 'org.apache.hadoop.hive.ql.
>    exec.FetchOperator:getRecordReader:FetchOperator.java:323',
>    
> 'org.apache.hadoop.hive.ql.exec.FetchOperator:getNextRow:FetchOperator.java:445',
>    '*java.lang.UnsupportedOperationException:parquet.column.values.
>    dictionary.PlainValuesDictionary$PlainLongDictionary:47:11',
>    'parquet.column.Dictionary:decodeToBinary:Dictionary.java:44', '
>    org.apache.hadoop.hive.ql.io.parquet.convert.ETypeConverter$
>    BinaryConverter:setDictionary:ETypeConverter.java:227',
>    'parquet.column.impl.ColumnReaderImpl:<init>:ColumnReaderImpl.java:339',
>    
> 'parquet.column.impl.ColumnReadStoreImpl:newMemColumnReader:ColumnReadStoreImpl.java:66',
>    
> 'parquet.column.impl.ColumnReadStoreImpl:getColumnReader:ColumnReadStoreImpl.java:61',
>    'parquet.io.RecordReaderImplementation:<init>:
>    RecordReaderImplementation.java:270', 'parquet.io.MessageColumnIO$1:
>    visit:MessageColumnIO.java:134', 'parquet.io.MessageColumnIO$1:
>    visit:MessageColumnIO.java:99', 'parquet.filter2.compat.
>    FilterCompat$NoOpFilter:accept:FilterCompat.java:154',
>    'parquet.io.MessageColumnIO:getRecordReader:MessageColumnIO.java:99',
>    'parquet.hadoop.InternalParquetRecordReader:checkRead:
>    InternalParquetRecordReader.java:137', 'parquet.hadoop.
>    
> InternalParquetRecordReader:nextKeyValue:InternalParquetRecordReader.java:208'],
>    statusCode=3), results=None, hasMoreRows=None)
>
> ls that something to do with hive? or a parquet error as such ?  I have
> posted it in both these groups, but I am afraid it mustnt be parquet
> because data is showing fine (viewed it in hue).
>
> Has anyone experienced similar errors before? Kindly let me know..
>
> Awaiting Your Reply,
>
> Thanks
> Sumit
>

Reply via email to