Hi, Glue catalog works just like any other Iceberg catalogs, by configuring related Iceberg catalog properties at engine start time. Within the Iceberg project provided integrations, you can use Spark, Flink, Hive by installing their respective runtime jars. Those jars package with the iceberg-aws module that contains the GlueCatalog implementation. Platforms like EMR automatically package the AWS v2 SDK dependencies for you, but if you do not use EMR, you will need to include additional AWS v2 SDKs. You can pick individual SDK clients, or use the bundle <https://mvnrepository.com/artifact/software.amazon.awssdk/bundle>. I also have this personal repo <https://github.com/jackye1995/aws-sdk-java-v2-analytics-bundle> as an example for a smaller size bundle.
For Trino/Presto-flavor engines, yes you can use Athena, and also any Trino offering in EMR through the Glue catalog connection: https://trino.io/docs/current/connector/iceberg.html#glue-catalog In addition, you can also use pyiceberg to connect to any Python engines and libraries: https://py.iceberg.apache.org/configuration/#glue-catalog Please let me know if you have any questions. Best, Jack Ye On Tue, Jun 20, 2023 at 9:02 AM Awasthi, Somesh <soawas...@informatica.com.invalid> wrote: > Hi Iceberg team, > > > > We want to know about how to perform read/write operation for iceberg > table present AWS Glue Catalog. > > > > As of now we know there is two approaches which we know to connect with > iceberg table. > > > > *Option1 JDBC:* Connect with Athena JDBC Driver > > > > *Option2 Spark:* Connect with Apache Iceberg Spark Runtime > > > > Could you please let me know is there any other solution other than SPARK > and JDBC API to connect with iceberg table for read/write operations. > > > > If there is an alternate way to perform read/write operation for iceberg > table present AWS Glue Catalog so please provide me steps in details to > implement it in sample program. > > > > Thanks, > > Somesh > > >