kazdy created HUDI-5578:
---------------------------

             Summary: Upgrade base docker image for java 8
                 Key: HUDI-5578
                 URL: https://issues.apache.org/jira/browse/HUDI-5578
             Project: Apache Hudi
          Issue Type: Improvement
          Components: dev-experience
            Reporter: kazdy
            Assignee: kazdy
             Fix For: 1.0.0


The existing Hudi docker image uses a pretty old Debian distribution, therefore 
it comes with Python 3.5.3. When using it with Spark 3.2.1,  running "pyspark" 
command in container shell results in below error:
{code:java}
root@e9fb3f81bdc9:/opt# pyspark
Python 3.5.3 (default, Nov  4 2021, 15:29:10) 
[GCC 6.3.0 20170516] on linux
Type "help", "copyright", "credits" or "license" for more information.
Traceback (most recent call last):
  File "/opt/spark/python/pyspark/shell.py", line 29, in <module>
    from pyspark.context import SparkContext
  File "/opt/spark/python/pyspark/__init__.py", line 53, in <module>
    from pyspark.rdd import RDD, RDDBarrier
  File "/opt/spark/python/pyspark/rdd.py", line 48, in <module>
    from pyspark.traceback_utils import SCCallSiteSync
  File "/opt/spark/python/pyspark/traceback_utils.py", line 23, in <module>
    CallSite = namedtuple("CallSite", "function file linenum")
  File "/opt/spark/python/pyspark/serializers.py", line 390, in namedtuple
    for k, v in _old_namedtuple_kwdefaults.items():
AttributeError: 'NoneType' object has no attribute 'items' {code}
The image I used was: 
[https://hub.docker.com/r/apachehudi/hudi-hadoop_3.1.0-hive_3.1.2-sparkadhoc_3.2.1/tags]

Spark 3.2.1 requires Python 3.6+, Spark 3.3 requires Python 3.7+.

Base image for Java 8 uses openjdk:8u212-jdk-slim-stretch.
The goal is to upgrade it to 
[8u342-jdk-slim-bullseye.|https://hub.docker.com/layers/library/openjdk/8u342-jdk-slim-bullseye/images/sha256-ecb89bb055c1ee4db9da38713b953f6daafefe575c77c6439eabbb85e3168402?context=explore]
This will also sync with existing java 11 base image distro.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to