I guess there are some confusions here between the metastore and the
actual Hive database. Spark (as well as Apache Hive) requires two
databases for Hive DB operations. Metastore is used for storing
metadata only (e.g., schema info), whereas the actual Hive database,
accessible through Thrift
Hello Artemis, Understand, if i gave hive metastore uri to anyone to connect
using pyspark. the port 9083 is open for anyone without authentication feature.
The only way pyspark able to connect to hive is through 9083 and not through
port 1.
On Friday, October 21, 2022 at 04:06:38 A
By default, Spark uses Apache Derby (running in embedded mode with store
content defined in local files) for hosting the Hive metastore. You can
externalize the metastore on a JDBC-compliant database (e.g.,
PostgreSQL) and use the database authentication provided by the
database. The JDBC con