I am trying to run a hive query from Spark code using HiveContext object. It was running fine earlier but since the Apache Sentry has been set installed the process is failing with this exception :
*org.apache.hadoop.security.AccessControlException: Permission denied: user=kakn, access=READ_EXECUTE, inode="/user/hive/warehouse/rt_freewheel_mastering.db/digital_profile_cluster_in":hive:hive:drwxrwx--t* I have pasted the full stack trace at the end of this post. My username "kakn" is a registered user with Sentry. I know that Spark takes all the configurations from hive-site.xml to execute the hql, so I added a few Sentry specific properties but seem to have no effect. I have attached the hive-site.xml *<property>* * <name>hive.security.authorization.task.factory</name>* * <value>org.apache.sentry.binding.hive.SentryHiveAuthorizationTaskFactoryImpl</value>* * </property>* * <property>* * <name>hive.metastore.pre.event.listeners</name>* * <value>org.apache.sentry.binding.metastore.MetastoreAuthzBinding</value>* * <description>list of comma seperated listeners for metastore events.</description>* * </property>* * <property>* * <name>hive.warehouse.subdir.inherit.perms</name>* * <value>true</value>* * </property>* *org.apache.hadoop.security.AccessControlException: Permission denied: user=kakn, access=READ_EXECUTE, inode="/user/hive/warehouse/rt_freewheel_mastering.db/digital_profile_cluster_in":hive:hive:drwxrwx--t* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:151)* * at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6287)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6269)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPathAccess(FSNamesystem.java:6194)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListingInt(FSNamesystem.java:4793)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:4755)* * at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:800)* * at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getListing(AuthorizationProviderProxyClientProtocol.java:310)* * at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:606)* * at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)* * at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:587)* * at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1026)* * at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)* * at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)* * at java.security.AccessController.doPrivileged(Native Method)* * at javax.security.auth.Subject.doAs(Subject.java:415)* * at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)* * at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)* * at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)* * at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)* * at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)* * at java.lang.reflect.Constructor.newInstance(Constructor.java:526)* * at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)* * at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)* * at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1895)* * at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1876)* * at org.apache.hadoop.hdfs.DistributedFileSystem.listStatusInternal(DistributedFileSystem.java:654)* * at org.apache.hadoop.hdfs.DistributedFileSystem.access$600(DistributedFileSystem.java:104)* * at org.apache.hadoop.hdfs.DistributedFileSystem$14.doCall(DistributedFileSystem.java:716)* * at org.apache.hadoop.hdfs.DistributedFileSystem$14.doCall(DistributedFileSystem.java:712)* * at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)* * at org.apache.hadoop.hdfs.DistributedFileSystem.listStatus(DistributedFileSystem.java:712)* * at org.apache.spark.sql.parquet.ParquetTypesConverter$.readMetaData(ParquetTypes.scala:440)* * at org.apache.spark.sql.parquet.ParquetTypesConverter$.readSchemaFromFile(ParquetTypes.scala:477)* * at org.apache.spark.sql.parquet.ParquetRelation.<init>(ParquetRelation.scala:65)* * at org.apache.spark.sql.SQLContext.parquetFile(SQLContext.scala:165)* * at org.apache.spark.sql.hive.HiveStrategies$ParquetConversion$.apply(HiveStrategies.scala:149)* * at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)* * at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)* * at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)* * at org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)* * at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:418)* * at org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:416)* * at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:422)* * at org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:422)* * at org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:425)* * at org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:425)* * at org.apache.spark.sql.SchemaRDD.getDependencies(SchemaRDD.scala:127)* * at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:192)* * at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:190)* * at scala.Option.getOrElse(Option.scala:120)* * at org.apache.spark.rdd.RDD.dependencies(RDD.scala:190)* * at org.apache.spark.rdd.RDD.firstParent(RDD.scala:1239)* * at org.apache.spark.sql.SchemaRDD.getPartitions(SchemaRDD.scala:122)* * at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205)* * at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:203)* * at scala.Option.getOrElse(Option.scala:120)* * at org.apache.spark.rdd.RDD.partitions(RDD.scala:203)* * at org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28)* * at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205)* * at org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:203)* * at scala.Option.getOrElse(Option.scala:120)* * at org.apache.spark.rdd.RDD.partitions(RDD.scala:203)* * at org.apache.spark.ShuffleDependency.<init>(Dependency.scala:79)* * at org.apache.spark.rdd.ShuffledRDD.getDependencies(ShuffledRDD.scala:80)* * at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:192)* * at org.apache.spark.rdd.RDD$$anonfun$dependencies$2.apply(RDD.scala:190)* * at scala.Option.getOrElse(Option.scala:120)* * at org.apache.spark.rdd.RDD.dependencies(RDD.scala:190)* * at org.apache.spark.scheduler.DAGScheduler.visit$1(DAGScheduler.scala:301)* * at org.apache.spark.scheduler.DAGScheduler.getParentStages(DAGScheduler.scala:313)* * at org.apache.spark.scheduler.DAGScheduler.newStage(DAGScheduler.scala:247)* * at org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:734)* * at org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1389)* * at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)* * at akka.actor.ActorCell.invoke(ActorCell.scala:456)* * at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)* * at akka.dispatch.Mailbox.run(Mailbox.scala:219)* * at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)* * at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)* * at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)* * at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)* * at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)* *Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=kakn, access=READ_EXECUTE, inode="/user/hive/warehouse/rt_freewheel_mastering.db/digital_profile_cluster_in":hive:hive:drwxrwx--t* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)* * at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:151)* * at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6287)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6269)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPathAccess(FSNamesystem.java:6194)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListingInt(FSNamesystem.java:4793)* * at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getListing(FSNamesystem.java:4755)* * at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getListing(NameNodeRpcServer.java:800)* * at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.getListing(AuthorizationProviderProxyClientProtocol.java:310)* * at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getListing(ClientNamenodeProtocolServerSideTranslatorPB.java:606)* * at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)* * at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:587)* * at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1026)* * at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)* * at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)* * at java.security.AccessController.doPrivileged(Native Method)* * at javax.security.auth.Subject.doAs(Subject.java:415)* * at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1642)* * at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)* * at org.apache.hadoop.ipc.Client.call(Client.java:1411)* * at org.apache.hadoop.ipc.Client.call(Client.java:1364)* * at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)* * at com.sun.proxy.$Proxy14.getListing(Unknown Source)* * at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:546)* * at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)* * at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)* * at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)* * at java.lang.reflect.Method.invoke(Method.java:606)* * at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)* * at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)* * at com.sun.proxy.$Proxy15.getListing(Unknown Source)* * at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1893)*
<?xml version="1.0" encoding="UTF-8"?> <!--Autogenerated by Cloudera Manager--> <configuration> <property> <name>hive.metastore.local</name> <value>false</value> </property> <property> <name>hive.metastore.uris</name> <value>thrift://blah.blah.com:9083</value> </property> <property> <name>hive.metastore.client.socket.timeout</name> <value>1800</value> </property> <property> <name>hive.metastore.warehouse.dir</name> <value>/user/hive/warehouse</value> </property> <property> <name>hive.warehouse.subdir.inherit.perms</name> <value>true</value> </property> <!--'mapred.reduce.tasks', originally set to '-1' (non-final), is overridden below by a safety valve--> <property> <name>hive.exec.reducers.bytes.per.reducer</name> <value>1073741824</value> </property> <property> <name>hive.exec.copyfile.maxsize</name> <value>33554432</value> </property> <property> <name>hive.exec.reducers.max</name> <value>999</value> </property> <property> <name>hive.metastore.execute.setugi</name> <value>true</value> </property> <property> <name>hive.support.concurrency</name> <value>true</value> </property> <property> <name>hive.zookeeper.quorum</name> <value>blah.blah.com,blah.blah.com,blah.blah.com</value> </property> <property> <name>hive.zookeeper.client.port</name> <value>2181</value> </property> <property> <name>hbase.zookeeper.quorum</name> <value>blah.blah.com,blah.blah.com,blah.blah.com</value> </property> <property> <name>hbase.zookeeper.property.clientPort</name> <value>2181</value> </property> <property> <name>hive.zookeeper.namespace</name> <value>hive_zookeeper_namespace_hive1</value> </property> <property> <name>hive.cluster.delegation.token.store.class</name> <value>org.apache.hadoop.hive.thrift.MemoryTokenStore</value> </property> <property> <name>hive.server2.enable.doAs</name> <value>false</value> </property> <property> <name>hive.server2.use.SSL</name> <value>false</value> </property> <property> <name>hive.merge.mapfiles</name> <value>true</value> </property> <property> <name>hive.merge.mapredfiles</name> <value>true</value> </property> <property> <name>hive.merge.size.per.task</name> <value>1024000000</value> </property> <property> <name>hive.merge.smallfiles.avgsize</name> <value>512000000</value> </property> <property> <name>hive.exec.compress.output</name> <value>true</value> </property> <property> <name>hive.exec.compress.intermediate</name> <value>true</value> </property> <property> <name>mapred.output.compression.codec</name> <value>org.apache.hadoop.io.compress.SnappyCodec</value> </property> <property> <name>mapred.output.compression.type</name> <value>BLOCK</value> </property> <property> <name>avro.output.codec</name> <value>snappy</value> </property> <property> <name>hive.auto.convert.join</name> <value>true</value> </property> <property> <name>hive.exec.parallel</name> <value>false</value> </property> <property> <name>hive.mapjoin.smalltable.filesize</name> <value>50000000</value> </property> <property> <name>hive.exec.parallel.thread.number</name> <value>12</value> </property> <property> <name>mapred.reduce.tasks</name> <value>-1</value> </property> <property> <name>hive.mapred.reduce.tasks.speculative.execution</name> <value>false</value> </property> <property> <name> hive.map.aggr.hash.min.reduction</name> <value>0.25</value> </property> <property> <name>hive.default.fileformat</name> <value>PARQUET</value> </property> <property> <name>parquet.compression</name> <value>SNAPPY</value> </property> <property> <name>hive.security.authorization.task.factory</name> <value>org.apache.sentry.binding.hive.SentryHiveAuthorizationTaskFactoryImpl</value> </property> <property> <name>hive.metastore.pre.event.listeners</name> <value>org.apache.sentry.binding.metastore.MetastoreAuthzBinding</value> <description>list of comma seperated listeners for metastore events.</description> </property> <property> <name>hive.warehouse.subdir.inherit.perms</name> <value>true</value> </property> </configuration>
--------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org