This is a bug in the HadoopInputSplit. It does not follow the general class loading rules in Flink. I think it is pretty straightforward to fix, I'll give it a quick shot...
Can you send me the entire stack trace (where the serialization call comes from) to verify this? On Wed, May 20, 2015 at 12:03 PM, Flavio Pompermaier <pomperma...@okkam.it> wrote: > Now I'm able to run the job but I get another exception..this time it > seems that Flink it's not able to split my Parquet file: > > Caused by: java.lang.ClassNotFoundException: > parquet.hadoop.ParquetInputSplit > at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > at java.lang.Class.forName0(Native Method) > at java.lang.Class.forName(Class.java:190) > at > org.apache.flink.api.java.hadoop.mapreduce.wrapper.HadoopInputSplit.readObject(HadoopInputSplit.java:104) > > I checked the jar and that class is present in my "fat" jar. > What should I do now? > > > On Wed, May 20, 2015 at 10:57 AM, Flavio Pompermaier <pomperma...@okkam.it > > wrote: > >> Yes it could be that the jar classes and those on the cluster are not >> aligned for some days..Now I'll recompile both sides and if I still have >> the error I will change line 42 as you suggested. >> Tanks Max >> >> On Wed, May 20, 2015 at 10:53 AM, Maximilian Michels <m...@apache.org> >> wrote: >> >>> Hi Flavio, >>> >>> It would be helpful, if we knew which class could not be found. In the >>> ClosureCleaner, can you change line 42 to include the class name in the >>> error message? Like in this example: >>> >>> private static ClassReader getClassReader(Class<?> cls) { >>> String className = cls.getName().replaceFirst("^.*\\.", "") + ".class"; >>> try { >>> return new ClassReader(cls.getResourceAsStream(className)); >>> } catch (IOException e) { >>> throw new RuntimeException("Could not create ClassReader for class " >>> + cls.getName() + ":" + e); >>> } >>> } >>> >>> Could it be that you're running an old job on the latest snapshot >>> version? This could cause class-related problems... >>> >>> Cheers, >>> Max >>> >>> On Wed, May 20, 2015 at 9:41 AM, Flavio Pompermaier < >>> pomperma...@okkam.it> wrote: >>> >>>> Any insight about this..? >>>> >>>> On Tue, May 19, 2015 at 12:49 PM, Flavio Pompermaier < >>>> pomperma...@okkam.it> wrote: >>>> >>>>> Hi to all, >>>>> >>>>> I tried to run my job on a brand new Flink cluster (0.9-SNAPSHOT) from >>>>> the web client UI using the shading strategy of the quickstart example but >>>>> I get this exception: >>>>> >>>>> Caused by: java.lang.RuntimeException: Could not create ClassReader: >>>>> java.io.IOException: Class not found >>>>> at >>>>> org.apache.flink.api.java.ClosureCleaner.getClassReader(ClosureCleaner.java:42) >>>>> at >>>>> org.apache.flink.api.java.ClosureCleaner.cleanThis0(ClosureCleaner.java:67) >>>>> at >>>>> org.apache.flink.api.java.ClosureCleaner.clean(ClosureCleaner.java:54) >>>>> >>>>> It seems that it cannot find some kryo class..how do I fix this? this >>>>> is my shade plugin section of pom.xml: >>>>> >>>>> >>>>> <plugin> >>>>> <groupId>org.apache.maven.plugins</groupId> >>>>> <artifactId>maven-shade-plugin</artifactId> >>>>> <version>1.4</version> >>>>> <executions> >>>>> <execution> >>>>> <phase>package</phase> >>>>> <goals> >>>>> <goal>shade</goal> >>>>> </goals> >>>>> <configuration> >>>>> <artifactSet> >>>>> <excludes> >>>>> <!-- This list contains all dependencies of flink-dist Everything >>>>> else will be packaged into the fat-jar --> >>>>> <exclude>org.apache.flink:flink-shaded-*</exclude> >>>>> <exclude>org.apache.flink:flink-core</exclude> >>>>> <exclude>org.apache.flink:flink-java</exclude> >>>>> <exclude>org.apache.flink:flink-scala</exclude> >>>>> <exclude>org.apache.flink:flink-runtime</exclude> >>>>> <exclude>org.apache.flink:flink-optimizer</exclude> >>>>> <exclude>org.apache.flink:flink-clients</exclude> >>>>> <exclude>org.apache.flink:flink-spargel</exclude> >>>>> <exclude>org.apache.flink:flink-avro</exclude> >>>>> <exclude>org.apache.flink:flink-java-examples</exclude> >>>>> <exclude>org.apache.flink:flink-scala-examples</exclude> >>>>> <exclude>org.apache.flink:flink-streaming-examples</exclude> >>>>> <exclude>org.apache.flink:flink-streaming-core</exclude> >>>>> >>>>> <!-- Also exclude very big transitive dependencies of Flink WARNING: >>>>> You have to remove these excludes if your code relies on other >>>>> versions of >>>>> these dependencies. --> >>>>> <exclude>org.scala-lang:scala-library</exclude> >>>>> <exclude>org.scala-lang:scala-compiler</exclude> >>>>> <exclude>org.scala-lang:scala-reflect</exclude> >>>>> <exclude>com.amazonaws:aws-java-sdk</exclude> >>>>> <exclude>com.typesafe.akka:akka-actor_*</exclude> >>>>> <exclude>com.typesafe.akka:akka-remote_*</exclude> >>>>> <exclude>com.typesafe.akka:akka-slf4j_*</exclude> >>>>> <exclude>io.netty:netty-all</exclude> >>>>> <exclude>io.netty:netty</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-server</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-continuation</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-http</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-io</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-util</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-security</exclude> >>>>> <exclude>org.eclipse.jetty:jetty-servlet</exclude> >>>>> <exclude>commons-fileupload:commons-fileupload</exclude> >>>>> <exclude>org.apache.avro:avro</exclude> >>>>> <exclude>commons-collections:commons-collections</exclude> >>>>> <exclude>org.codehaus.jackson:jackson-core-asl</exclude> >>>>> <exclude>org.codehaus.jackson:jackson-mapper-asl</exclude> >>>>> <exclude>com.thoughtworks.paranamer:paranamer</exclude> >>>>> <exclude>org.xerial.snappy:snappy-java</exclude> >>>>> <exclude>org.apache.commons:commons-compress</exclude> >>>>> <exclude>org.tukaani:xz</exclude> >>>>> <exclude>com.esotericsoftware.kryo:kryo</exclude> >>>>> <exclude>com.esotericsoftware.minlog:minlog</exclude> >>>>> <exclude>org.objenesis:objenesis</exclude> >>>>> <exclude>com.twitter:chill_*</exclude> >>>>> <exclude>com.twitter:chill-java</exclude> >>>>> <exclude>com.twitter:chill-avro_*</exclude> >>>>> <exclude>com.twitter:chill-bijection_*</exclude> >>>>> <exclude>com.twitter:bijection-core_*</exclude> >>>>> <exclude>com.twitter:bijection-avro_*</exclude> >>>>> <exclude>com.twitter:chill-protobuf</exclude> >>>>> <exclude>com.google.protobuf:protobuf-java</exclude> >>>>> <exclude>com.twitter:chill-thrift</exclude> >>>>> <exclude>org.apache.thrift:libthrift</exclude> >>>>> <exclude>commons-lang:commons-lang</exclude> >>>>> <exclude>junit:junit</exclude> >>>>> <exclude>de.javakaffee:kryo-serializers</exclude> >>>>> <exclude>joda-time:joda-time</exclude> >>>>> <exclude>org.apache.commons:commons-lang3</exclude> >>>>> <exclude>org.slf4j:slf4j-api</exclude> >>>>> <exclude>org.slf4j:slf4j-log4j12</exclude> >>>>> <exclude>log4j:log4j</exclude> >>>>> <exclude>org.apache.commons:commons-math</exclude> >>>>> <exclude>org.apache.sling:org.apache.sling.commons.json</exclude> >>>>> <exclude>commons-logging:commons-logging</exclude> >>>>> <exclude>org.apache.httpcomponents:httpclient</exclude> >>>>> <exclude>org.apache.httpcomponents:httpcore</exclude> >>>>> <exclude>commons-codec:commons-codec</exclude> >>>>> <exclude>com.fasterxml.jackson.core:jackson-core</exclude> >>>>> <exclude>com.fasterxml.jackson.core:jackson-databind</exclude> >>>>> <exclude>com.fasterxml.jackson.core:jackson-annotations</exclude> >>>>> <exclude>org.codehaus.jettison:jettison</exclude> >>>>> <exclude>stax:stax-api</exclude> >>>>> <exclude>com.typesafe:config</exclude> >>>>> <exclude>org.uncommons.maths:uncommons-maths</exclude> >>>>> <exclude>com.github.scopt:scopt_*</exclude> >>>>> <exclude>org.mortbay.jetty:servlet-api</exclude> >>>>> <exclude>commons-io:commons-io</exclude> >>>>> <exclude>commons-cli:commons-cli</exclude> >>>>> </excludes> >>>>> </artifactSet> >>>>> <filters> >>>>> <filter> >>>>> <artifact>org.apache.flink:*</artifact> >>>>> <excludes> >>>>> <exclude>org/apache/flink/shaded/**</exclude> >>>>> <exclude>web-docs/**</exclude> >>>>> </excludes> >>>>> </filter> >>>>> </filters> >>>>> <createDependencyReducedPom>false</createDependencyReducedPom> >>>>> <finalName>XXXX</finalName> >>>>> <transformers> >>>>> <!-- add Main-Class to manifest file --> >>>>> <transformer >>>>> >>>>> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> >>>>> <manifestEntries> >>>>> <Main-Class>XXX</Main-Class> >>>>> </manifestEntries> >>>>> </transformer> >>>>> </transformers> >>>>> </configuration> >>>>> </execution> >>>>> </executions> >>>>> </plugin> >>>>> >>>> >>>> >>> >> >> > >