Hi,

Glue catalog works just like any other Iceberg catalogs, by configuring
related Iceberg catalog properties at engine start time. Within the Iceberg
project provided integrations, you can use Spark, Flink, Hive by installing
their respective runtime jars. Those jars package with the iceberg-aws
module that contains the GlueCatalog implementation. Platforms like EMR
automatically package the AWS v2 SDK dependencies for you, but if you do
not use EMR, you will need to include additional AWS v2 SDKs. You can pick
individual SDK clients, or use the bundle
<https://mvnrepository.com/artifact/software.amazon.awssdk/bundle>. I also
have this personal repo
<https://github.com/jackye1995/aws-sdk-java-v2-analytics-bundle> as an
example for a smaller size bundle.

For Trino/Presto-flavor engines, yes you can use Athena, and also any Trino
offering in EMR through the Glue catalog connection:
https://trino.io/docs/current/connector/iceberg.html#glue-catalog

In addition, you can also use pyiceberg to connect to any Python engines
and libraries: https://py.iceberg.apache.org/configuration/#glue-catalog

Please let me know if you have any questions.

Best,
Jack Ye







On Tue, Jun 20, 2023 at 9:02 AM Awasthi, Somesh
<soawas...@informatica.com.invalid> wrote:

> Hi Iceberg team,
>
>
>
> We want to know about how to perform read/write operation for iceberg
> table present AWS Glue Catalog.
>
>
>
> As of now we know there is two approaches which we know to connect with
> iceberg table.
>
>
>
> *Option1 JDBC:* Connect with Athena JDBC Driver
>
>
>
> *Option2 Spark:*  Connect with Apache Iceberg Spark Runtime
>
>
>
> Could you please let me know is there any other solution other than SPARK
> and JDBC API to connect with iceberg table for read/write operations.
>
>
>
> If there is an alternate way to perform read/write operation for iceberg
> table present AWS Glue Catalog so please provide me steps in details to
> implement it in sample program.
>
>
>
> Thanks,
>
> Somesh
>
>
>

Reply via email to