Hi Jacob, It’s a little bit of guesswork …
The disappearing records remind me a bit of a peculiarity of Oracle, that each (e.g. INSERT) statement is in an implicit transaction and hence needs to be committed. In Flink committing transaction happen together with the checkpoint cycle, i.e. this needs to be setup properly for your job. I work mostly with streaming API not table API, but I guess there the matter is just the same. For the database in use (instead of ‘default’), I think you can specify this in your JDBC connection string. Hope that helps 😊 Thias From: Jacob Rollings <jacobrolling...@gmail.com> Sent: Friday, November 8, 2024 7:39 AM To: user@flink.apache.org Subject: [External] Re: Flink table materialization ⚠EXTERNAL MESSAGE – CAUTION: Think Before You Click ⚠ After correcting the properties of the connector, now iam getting the error in the screenshot I have also attached the jars in the class path while launching flink sql cli. Full description of the usecase is in the first email in this loop. On Thu, Nov 7, 2024, 11:38 PM Jacob Rollings <jacobrolling...@gmail.com<mailto:jacobrolling...@gmail.com>> wrote: Added attachment of the error message. On Thu, Nov 7, 2024, 11:12 PM Jacob Rollings <jacobrolling...@gmail.com<mailto:jacobrolling...@gmail.com>> wrote: Hi, I want to make the tables created by Flink Table API/SQL durable and permanent. To achieve this, I am trying the following basic example using the JDBC Oracle connector. I have added both the Flink JDBC and Oracle JDBC drivers to the Flink lib directory. I am using the Flink SQL client to run the queries. While it successfully creates tables in the default-database, I don't see the actual tables in the Oracle database. Moreover, the tables created under the Flink default-database seem to exist only as long as the session is active. What steps should I take to ensure that the in-memory tables I work with during my Flink job are permanently stored in the database? The documentation mentions using Catalogs and Connectors to persist in-memory tables to a database. I want to store my tables and data in Oracle DB. However, I noticed that Flink supports only Hive, PostgreSQL, and MySQL catalogs. Does this mean my data will reside in Oracle while the metadata about the tables can be stored only in a Hive, PostgreSQL, or MySQL metastore? The documentation on this topic seems to cover only basic concepts and lacks complete examples on how to achieve this. Any pointers or detailed guidance would be greatly appreciated. Thanks. Diese Nachricht ist ausschliesslich für den Adressaten bestimmt und beinhaltet unter Umständen vertrauliche Mitteilungen. Da die Vertraulichkeit von e-Mail-Nachrichten nicht gewährleistet werden kann, übernehmen wir keine Haftung für die Gewährung der Vertraulichkeit und Unversehrtheit dieser Mitteilung. Bei irrtümlicher Zustellung bitten wir Sie um Benachrichtigung per e-Mail und um Löschung dieser Nachricht sowie eventueller Anhänge. Jegliche unberechtigte Verwendung oder Verbreitung dieser Informationen ist streng verboten. This message is intended only for the named recipient and may contain confidential or privileged information. As the confidentiality of email communication cannot be guaranteed, we do not accept any responsibility for the confidentiality and the intactness of this message. If you have received it in error, please advise the sender by return e-mail and delete this message and any attachments. Any unauthorised use or dissemination of this information is strictly prohibited.