[ 
https://issues.apache.org/jira/browse/SPARK-53226?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dongjoon Hyun updated SPARK-53226:
----------------------------------
    Description: 
{code}
$ bin/spark-shell
WARNING: Using incubator modules: jdk.incubator.vector
WARNING: package sun.security.action not in java.base
Using Spark's default log4j profile: org/apache/spark/log4j2-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
setLogLevel(newLevel).
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 4.1.0-preview1
      /_/

Using Scala version 2.13.16 (OpenJDK 64-Bit Server VM, Java 25)
Type in expressions to have them evaluated.
Type :help for more information.
25/08/09 13:21:21 WARN NativeCodeLoader: Unable to load native-hadoop library 
for your platform... using builtin-java classes where applicable
WARNING: A terminally deprecated method in sun.misc.Unsafe has been called
WARNING: sun.misc.Unsafe::allocateMemory has been called by 
io.netty.util.internal.PlatformDependent0$2 
(file:/Users/dongjoon/APACHE/spark-4.1.0-preview1-bin-hadoop3/jars/netty-common-4.1.122.Final.jar)
WARNING: Please consider reporting this to the maintainers of class 
io.netty.util.internal.PlatformDependent0$2
WARNING: sun.misc.Unsafe::allocateMemory will be removed in a future release
Spark context Web UI available at http://localhost:4040
Spark context available as 'sc' (master = local[*], app id = 
local-1754770881328).
Spark session available as 'spark'.

scala> :paste
// Entering paste mode (ctrl-D to finish)

var array = new Array[Int](5)
val broadcastArray = sc.broadcast(array)
sc.parallelize(0 to 4).map(x => broadcastArray.value(x)).collect()
array(0) = 5
sc.parallelize(0 to 4).map(x => broadcastArray.value(x)).collect()

// Exiting paste mode... now interpreting.
java.lang.InternalError: java.lang.IllegalAccessException: final field has no 
write access: $Lambda/0x000000f001979948.arg$1/putField, from class 
java.lang.Object (module java.base)
  at 
java.base/jdk.internal.reflect.MethodHandleAccessorFactory.newFieldAccessor(MethodHandleAccessorFactory.java:208)
  at 
java.base/jdk.internal.reflect.ReflectionFactory.newFieldAccessor(ReflectionFactory.java:109)
  at 
java.base/java.lang.reflect.Field.acquireOverrideFieldAccessor(Field.java:1195)
  at java.base/java.lang.reflect.Field.getOverrideFieldAccessor(Field.java:1164)
  at java.base/java.lang.reflect.Field.set(Field.java:832)
  at 
org.apache.spark.util.ClosureCleaner$.setFieldAndIgnoreModifiers(ClosureCleaner.scala:564)
  at 
org.apache.spark.util.ClosureCleaner$.cleanupScalaReplClosure(ClosureCleaner.scala:432)
  at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:257)
  at 
org.apache.spark.util.SparkClosureCleaner$.clean(SparkClosureCleaner.scala:39)
  at org.apache.spark.SparkContext.clean(SparkContext.scala:2843)
  at org.apache.spark.rdd.RDD.$anonfun$map$1(RDD.scala:425)
  at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
  at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
  at org.apache.spark.rdd.RDD.withScope(RDD.scala:417)
  at org.apache.spark.rdd.RDD.map(RDD.scala:424)
  ... 46 elided
Caused by: java.lang.IllegalAccessException: final field has no write access: 
$Lambda/0x000000f001979948.arg$1/putField, from class java.lang.Object (module 
java.base)
  at 
java.base/java.lang.invoke.MemberName.makeAccessException(MemberName.java:889)
  at 
java.base/java.lang.invoke.MethodHandles$Lookup.unreflectField(MethodHandles.java:3441)
  at 
java.base/java.lang.invoke.MethodHandles$Lookup.unreflectSetter(MethodHandles.java:3432)
  at 
java.base/java.lang.invoke.MethodHandleImpl$1.unreflectField(MethodHandleImpl.java:1607)
  at 
java.base/jdk.internal.reflect.MethodHandleAccessorFactory.newFieldAccessor(MethodHandleAccessorFactory.java:186)
  ... 60 more
{code}

> Spark-Shell run fails with Java 25
> ----------------------------------
>
>                 Key: SPARK-53226
>                 URL: https://issues.apache.org/jira/browse/SPARK-53226
>             Project: Spark
>          Issue Type: Sub-task
>          Components: Spark Core
>    Affects Versions: 4.1.0
>            Reporter: Dongjoon Hyun
>            Priority: Major
>
> {code}
> $ bin/spark-shell
> WARNING: Using incubator modules: jdk.incubator.vector
> WARNING: package sun.security.action not in java.base
> Using Spark's default log4j profile: 
> org/apache/spark/log4j2-defaults.properties
> Setting default log level to "WARN".
> To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use 
> setLogLevel(newLevel).
> Welcome to
>       ____              __
>      / __/__  ___ _____/ /__
>     _\ \/ _ \/ _ `/ __/  '_/
>    /___/ .__/\_,_/_/ /_/\_\   version 4.1.0-preview1
>       /_/
> Using Scala version 2.13.16 (OpenJDK 64-Bit Server VM, Java 25)
> Type in expressions to have them evaluated.
> Type :help for more information.
> 25/08/09 13:21:21 WARN NativeCodeLoader: Unable to load native-hadoop library 
> for your platform... using builtin-java classes where applicable
> WARNING: A terminally deprecated method in sun.misc.Unsafe has been called
> WARNING: sun.misc.Unsafe::allocateMemory has been called by 
> io.netty.util.internal.PlatformDependent0$2 
> (file:/Users/dongjoon/APACHE/spark-4.1.0-preview1-bin-hadoop3/jars/netty-common-4.1.122.Final.jar)
> WARNING: Please consider reporting this to the maintainers of class 
> io.netty.util.internal.PlatformDependent0$2
> WARNING: sun.misc.Unsafe::allocateMemory will be removed in a future release
> Spark context Web UI available at http://localhost:4040
> Spark context available as 'sc' (master = local[*], app id = 
> local-1754770881328).
> Spark session available as 'spark'.
> scala> :paste
> // Entering paste mode (ctrl-D to finish)
> var array = new Array[Int](5)
> val broadcastArray = sc.broadcast(array)
> sc.parallelize(0 to 4).map(x => broadcastArray.value(x)).collect()
> array(0) = 5
> sc.parallelize(0 to 4).map(x => broadcastArray.value(x)).collect()
> // Exiting paste mode... now interpreting.
> java.lang.InternalError: java.lang.IllegalAccessException: final field has no 
> write access: $Lambda/0x000000f001979948.arg$1/putField, from class 
> java.lang.Object (module java.base)
>   at 
> java.base/jdk.internal.reflect.MethodHandleAccessorFactory.newFieldAccessor(MethodHandleAccessorFactory.java:208)
>   at 
> java.base/jdk.internal.reflect.ReflectionFactory.newFieldAccessor(ReflectionFactory.java:109)
>   at 
> java.base/java.lang.reflect.Field.acquireOverrideFieldAccessor(Field.java:1195)
>   at 
> java.base/java.lang.reflect.Field.getOverrideFieldAccessor(Field.java:1164)
>   at java.base/java.lang.reflect.Field.set(Field.java:832)
>   at 
> org.apache.spark.util.ClosureCleaner$.setFieldAndIgnoreModifiers(ClosureCleaner.scala:564)
>   at 
> org.apache.spark.util.ClosureCleaner$.cleanupScalaReplClosure(ClosureCleaner.scala:432)
>   at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:257)
>   at 
> org.apache.spark.util.SparkClosureCleaner$.clean(SparkClosureCleaner.scala:39)
>   at org.apache.spark.SparkContext.clean(SparkContext.scala:2843)
>   at org.apache.spark.rdd.RDD.$anonfun$map$1(RDD.scala:425)
>   at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>   at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>   at org.apache.spark.rdd.RDD.withScope(RDD.scala:417)
>   at org.apache.spark.rdd.RDD.map(RDD.scala:424)
>   ... 46 elided
> Caused by: java.lang.IllegalAccessException: final field has no write access: 
> $Lambda/0x000000f001979948.arg$1/putField, from class java.lang.Object 
> (module java.base)
>   at 
> java.base/java.lang.invoke.MemberName.makeAccessException(MemberName.java:889)
>   at 
> java.base/java.lang.invoke.MethodHandles$Lookup.unreflectField(MethodHandles.java:3441)
>   at 
> java.base/java.lang.invoke.MethodHandles$Lookup.unreflectSetter(MethodHandles.java:3432)
>   at 
> java.base/java.lang.invoke.MethodHandleImpl$1.unreflectField(MethodHandleImpl.java:1607)
>   at 
> java.base/jdk.internal.reflect.MethodHandleAccessorFactory.newFieldAccessor(MethodHandleAccessorFactory.java:186)
>   ... 60 more
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to