-1 non binding. We are finding failures in astro sdk DAGs after using the
databricks and google RC.
We are working on creating PRs for the fixes.

FAILED
tests_integration/databases/databricks_tests/test_delta.py::test_delta_run_sql
- AttributeError: 'list' object has no attribute 'asDict'
FAILED
tests_integration/databases/databricks_tests/test_delta.py::test_delta_run_sql_with_parameters
- AttributeError: 'list' object has no attribute 'asDict'
FAILED
tests_integration/databases/databricks_tests/test_delta.py::test_delta_create_table_with_columns[delta]
- AssertionError: assert ['id', 'int', None] == Row(col_name='id',
data_type='int', comment=None)
  Full diff:
  - Row(col_name='id', data_type='int', comment=None)
  + ['id', 'int', None]
ERROR
tests_integration/databases/databricks_tests/test_delta.py::test_create_table_from_select_statement[delta]
- airflow.exceptions.AirflowException: Databricks job failed. Job info
***'job_id': 438976959785934, 'run_id': 268958235083047,
'creator_user_name': 'phani.ku...@astronomer.io', 'number_in_job':
268958235083047, 'state': ***'life_cycle_state': 'TERMINATED',
'result_state': 'FAILED', 'state_message': 'Workload failed, see run output
for details', 'user_cancelled_or_timedout': False***, 'task':
***'spark_python_task': ***'python_file':
'dbfs:/mnt/pyscripts/load_file__tmp_en0ra8imcxfee4sp8b9y4hzqj4bjy5y2fvgpo9gv0s3pzvmzqv1tfaza9.py'***,
'cluster_spec': ***'existing_cluster_id': '***'***, 'cluster_instance':
***'cluster_id': '***', 'spark_context_id': '4902558347078657686'***,
'start_time': 1702436831591, 'setup_duration': 1000, 'execution_duration':
12000, 'cleanup_duration': 0, 'end_time': 1702436844777, 'run_name':
'Untitled', 'run_page_url': '
https://dbc-9c390870-65ef.cloud.databricks.com/?o=4256138892007661#job/438976959785934/run/268958235083047',
'run_type': 'SUBMIT_RUN', 'attempt_number': 0, 'format': 'SINGLE_TASK'***

On Wed, Dec 13, 2023 at 4:02 AM Jarek Potiuk <ja...@potiuk.com> wrote:

> Hey all,
>
> [Filling-in for Elad, who had no time to do it this time]
>
> I have just cut the new wave Airflow Providers packages. This email is
> calling a vote on the release,
> which will last for 24 hours - which means that it will end on December 13,
> 2023 22:00 PM UTC and until 3 binding +1 votes have been received.
>
> Following our processes, this is an accelerated vote taking into account
> that the RC1 version has been tested and it only contains incremental
> changes.
>
> Consider this my (binding) +1.
>
> This release contains fixes to issues found in rc1 in google, docker, odbc,
> databricks providers and the last (after scheduling for
> removal) daskexecutor provider release.
>
> Airflow Providers are available at:
> https://dist.apache.org/repos/dist/dev/airflow/providers/
>
> *apache-airflow-providers-<PROVIDER>-*.tar.gz* are the binary
>  Python "sdist" release - they are also official "sources" for the provider
> packages.
>
> *apache_airflow_providers_<PROVIDER>-*.whl are the binary
>  Python "wheel" release.
>
> The test procedure for PMC members is described in
>
> https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-the-release-candidate-by-pmc-members
>
> The test procedure for and Contributors who would like to test this RC is
> described in:
>
> https://github.com/apache/airflow/blob/main/dev/README_RELEASE_PROVIDER_PACKAGES.md#verify-the-release-candidate-by-contributors
>
> Public keys are available at:
> https://dist.apache.org/repos/dist/release/airflow/KEYS
>
> Please vote accordingly:
>
> [ ] +1 approve
> [ ] +0 no opinion
> [ ] -1 disapprove with the reason
>
>
> Only votes from PMC members are binding, but members of the community are
> encouraged to test the release and vote with "(non-binding)".
>
> Please note that the version number excludes the 'rcX' string.
> This will allow us to rename the artifact without modifying
> the artifact checksums when we actually release.
>
> The status of testing the providers by the community is kept here:
>
> https://github.com/apache/airflow/issues/36194
>
> You can find packages as well as detailed changelog following the below
> links:
>
> https://pypi.org/project/apache-airflow-providers-daskexecutor/1.1.1rc2/
> https://pypi.org/project/apache-airflow-providers-google/10.13.0rc2/
> https://pypi.org/project/apache-airflow-providers-databricks/5.1.0rc2/
> https://pypi.org/project/apache-airflow-providers-docker/3.9.0rc2/
> https://pypi.org/project/apache-airflow-providers-odbc/4.3.0rc2/
>
> Cheers,
>
> J.
>

Reply via email to