Sorry for the late reply Rog, been kinda busy lately.

Please look into the schema evolution of Avro. Confluent has an excellent
article on this: https://docs.confluent.io/current/schema-registry/avro.html

Could you try again with optional fields? e.g. "type": ["null", "array"].

Since the names are different, I would expect the default value (or even an
exception). If you do a cat on the Avro file, you can see that the original
schema is in the header of the file. The B field is not there the record,
so the reader field is not compatible, so it won't work. I'll check if we
can come up with a more meaningful exception.

Cheers, Fokko



Op vr 17 jan. 2020 om 17:02 schreef roger peppe <rogpe...@gmail.com>:

>
>
> On Fri, 17 Jan 2020 at 13:35, Ryan Skraba <r...@skraba.com> wrote:
>
>> Hello!  I just created a JIRA for this as an improvement :D
>> https://issues.apache.org/jira/browse/AVRO-2689
>>
>> To check evolution, we'd probably want to specify the reader schema in
>> the GenericDatumReader created here:
>>
>> https://github.com/apache/avro/blob/master/lang/java/tools/src/main/java/org/apache/avro/tool/DataFileReadTool.java#L75
>>
>> The writer schema is automatically set when the DataFileStream is
>> created.  If we want to set a different reader schema (than the one
>> found in the file), it should be set by calling
>> reader.setExpected(readerSchema) just after the DataFileStream is
>> created.
>>
>
> Ah, that's a good pointer, thanks! I was looking for an appropriate
> constructor, but there didn't seem to be one.
>
>
>>
>> I think it's a pretty good idea -- it feels like we're seeing more
>> questions about schema evolution these days, so that would be a neat
>> way for a user to test (or to create reproducible scenarios for bug
>> reports).  If you're interested, feel free to take the JIRA!  I'd be
>> happy to help out.
>>
>
> So, I've had a go at it... see
> https://github.com/rogpeppe-contrib/avro/commit/1236e9d33207a11d557c1eb2a171972e085dfcf2
>
> I did the following to see if it was working ("avro" is my shell script
> wrapper around the avro-tools jar):
>
> % cat schema.avsc
> {
>   "name": "R",
>   "type": "record",
>   "fields": [
>     {
>       "name": "A",
>       "type": {
>         "type": "array",
>         "items": "int"
>       }
>     }
>   ]
> }
> % cat schema1.avsc
> {
>   "name": "R",
>   "type": "record",
>   "fields": [
>     {
>       "name": "B",
>       "type": "string",
>       "default": "hello"
>     }
>   ]
> }
> %
> AVRO_TOOLS_JAR=/home/rog/other/avro/lang/java/tools/target/avro-tools-1.10.0-SNAPSHOT.ja%
> avro random --count 1 --schema-file schema.avsc x.out
> % avro tojson x.out
> {"A":[-890831012,1123049230,302974832]}
> % cp schema.avsc schema1.avsc
> % avro tojson --reader-schema-file schema1.avsc x.out
> Exception in thread "main" java.lang.ClassCastException: class
> org.apache.avro.util.Utf8 cannot be cast to class java.util.Collection
> (org.apache.avro.util.Utf8 is in unnamed module of loader 'app';
> java.util.Collection is in module java.base of loader 'bootstrap')
> at
> org.apache.avro.generic.GenericDatumWriter.getArraySize(GenericDatumWriter.java:258)
> at
> org.apache.avro.generic.GenericDatumWriter.writeArray(GenericDatumWriter.java:228)
> at
> org.apache.avro.generic.GenericDatumWriter.writeWithoutConversion(GenericDatumWriter.java:136)
> at
> org.apache.avro.generic.GenericDatumWriter.write(GenericDatumWriter.java:82)
> at
> org.apache.avro.generic.GenericDatumWriter.writeField(GenericDatumWriter.java:206)
> at
> org.apache.avro.generic.GenericDatumWriter.writeRecord(GenericDatumWriter.java:195)
> at
> org.apache.avro.generic.GenericDatumWriter.writeWithoutConversion(GenericDatumWriter.java:130)
> at
> org.apache.avro.generic.GenericDatumWriter.write(GenericDatumWriter.java:82)
> at
> org.apache.avro.generic.GenericDatumWriter.write(GenericDatumWriter.java:72)
> at org.apache.avro.tool.DataFileReadTool.run(DataFileReadTool.java:99)
> at org.apache.avro.tool.Main.run(Main.java:66)
> at org.apache.avro.tool.Main.main(Main.java:55)
> %
>
> I am a bit clueless when it comes to interpreting that exception... sorry
> for the ignorance - this is the first Java code I've ever written!
> Any idea what's going on? This is maybe getting a bit too noisy for the
> list - feel to reply directly.
>
>   cheers,
>     rog.
>
>
>> Ryan
>>
>>
>> On Fri, Jan 17, 2020 at 2:22 PM roger peppe <rogpe...@gmail.com> wrote:
>> >
>> > On Thu, 16 Jan 2020 at 17:21, Ryan Skraba <r...@skraba.com> wrote:
>> >>
>> >> didn't find anything currently in the avro-tools that uses both
>> >> reader and writer schemas while deserializing data...  It should be a
>> >> pretty easy feature to add as an option to the DataFileReadTool
>> >> (a.k.a. tojson)!
>> >
>> >
>> > Thanks for that suggestion. I've been delving into that code a bit and
>> trying to understand what's going on.
>> >
>> > At the heart of it is this code:
>> >
>> >     GenericDatumReader<Object> reader = new GenericDatumReader<>();
>> >     try (DataFileStream<Object> streamReader = new
>> DataFileStream<>(inStream, reader)) {
>> >       Schema schema = streamReader.getSchema();
>> >       DatumWriter<Object> writer = new GenericDatumWriter<>(schema);
>> >       JsonEncoder encoder = EncoderFactory.get().jsonEncoder(schema,
>> out, pretty);
>> >
>> > I'm trying to work out where the best place to put the specific reader
>> schema (taken from a command line flag) might be.
>> >
>> > Would it be best to do it when creating the DatumReader (it looks like
>> there might be a way to create that with a generic writer schema and a
>> specific reader schema, although I can't quite see how to do that atm), or
>> when creating the DatumWriter?
>> > Or perhaps there's a better way?
>> >
>> > Thanks for any guidance.
>> >
>> >    cheers,
>> >     rog.
>> >>
>> >>
>> >> You are correct about running ./build.sh dist in the java directory --
>> >> it fails with JDK 11 (likely fixable:
>> >> https://issues.apache.org/jira/browse/MJAVADOC-562).
>> >>
>> >> You should probably do a simple mvn clean install instead and find the
>> >> jar in lang/java/tools/target/avro-tools-1.10.0-SNAPSHOT.jar.  That
>> >> should work with JDK11 without any problem (well-tested in the build).
>> >>
>> >> Best regards, Ryan
>> >>
>> >>
>> >>
>> >> On Thu, Jan 16, 2020 at 5:49 PM roger peppe <rogpe...@gmail.com>
>> wrote:
>> >> >
>> >> > Update: I tried running `build.sh dist` in `lang/java` and it failed
>> (at least, it looks like a failure message) after downloading a load of
>> Maven deps with the following errors:
>> https://gist.github.com/rogpeppe/df05d993254dc5082253a5ef5027e965
>> >> >
>> >> > Any hints on what I should do to build the avro-tools jar?
>> >> >
>> >> >   cheers,
>> >> >     rog.
>> >> >
>> >> > On Thu, 16 Jan 2020 at 16:45, roger peppe <rogpe...@gmail.com>
>> wrote:
>> >> >>
>> >> >>
>> >> >> On Thu, 16 Jan 2020 at 13:57, Ryan Skraba <r...@skraba.com> wrote:
>> >> >>>
>> >> >>> Hello!  Is it because you are using brew to install avro-tools?
>> I'm
>> >> >>> not entirely familiar with how it packages the command, but using a
>> >> >>> direct bash-like solution instead might solve this problem of
>> mixing
>> >> >>> stdout and stderr.  This could be the simplest (and right) solution
>> >> >>> for piping.
>> >> >>
>> >> >>
>> >> >> No, I downloaded the jar and am directly running it with "java -jar
>> ~/other/avro-tools-1.9.1.jar".
>> >> >> I'm using Ubuntu Linux 18.04 FWIW - the binary comes from Debian
>> package openjdk-11-jre-headless.
>> >> >>
>> >> >> I'm going to try compiling avro-tools myself to investigate but I'm
>> a total Java ignoramus - wish me luck!
>> >> >>
>> >> >>>
>> >> >>> alias avrotoolx='java -jar
>> >> >>>
>> ~/.m2/repository/org/apache/avro/avro-tools/1.9.1/avro-tools-1.9.1.jar'
>> >> >>> avrotoolx tojson x.out 2> /dev/null
>> >> >>>
>> >> >>> (As Fokko mentioned, the 2> /dev/null isn't even necessary -- the
>> >> >>> warnings and logs should not be piped along with the normal
>> content.)
>> >> >>>
>> >> >>> Otherwise, IIRC, there is no way to disable the first illegal
>> >> >>> reflective access warning when running in Java 9+, but you can
>> "fix"
>> >> >>> these module errors, and deactivate the NativeCodeLoader logs with
>> an
>> >> >>> explicit log4j.properties:
>> >> >>>
>> >> >>> java -Dlog4j.configuration=file:///tmp/log4j.properties --add-opens
>> >> >>> java.security.jgss/sun.security.krb5=ALL-UNNAMED -jar
>> >> >>>
>> ~/.m2/repository/org/apache/avro/avro-tools/1.9.1/avro-tools-1.9.1.jar
>> >> >>> tojson x.out
>> >> >>
>> >> >>
>> >> >> Thanks for that suggestion! I'm afraid I'm not familiar with log4j
>> properties files though. What do I need to put in /tmp/log4j.properties to
>> make this work?
>> >> >>
>> >> >>> None of that is particularly satisfactory, but it could be a
>> >> >>> workaround for your immediate use.
>> >> >>
>> >> >>
>> >> >> Yeah, not ideal, because if something goes wrong, stdout will be
>> corrupted, but at least some noise should go away :)
>> >> >>
>> >> >>> I'd also like to see a more unified experience with the CLI tool
>> for
>> >> >>> documentation and usage.  The current state requires a bit of Avro
>> >> >>> expertise to use, but it has some functions that would be pretty
>> >> >>> useful for a user working with Avro data.  I raised
>> >> >>> https://issues.apache.org/jira/browse/AVRO-2688 as an improvement.
>> >> >>>
>> >> >>> In my opinion, a schema compatibility tool would be a useful and
>> >> >>> welcome feature!
>> >> >>
>> >> >>
>> >> >> That would indeed be nice, but in the meantime, is there really
>> nothing in the avro-tools commands that uses a chosen schema to read a data
>> file written with some other schema? That would give me what I'm after
>> currently.
>> >> >>
>> >> >> Thanks again for the helpful response.
>> >> >>
>> >> >>    cheers,
>> >> >>      rog.
>> >> >>
>> >> >>>
>> >> >>> Best regards, Ryan
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> On Thu, Jan 16, 2020 at 12:25 PM roger peppe <rogpe...@gmail.com>
>> wrote:
>> >> >>> >
>> >> >>> > Hi Fokko,
>> >> >>> >
>> >> >>> > Thanks for your swift response!
>> >> >>> >
>> >> >>> > Stdout and stderr definitely seem to be merged on this platform
>> at least. Here's a sample:
>> >> >>> >
>> >> >>> > % avrotool random --count 1 --schema '"int"'  x.out
>> >> >>> > % avrotool tojson x.out > x.json
>> >> >>> > % cat x.json
>> >> >>> > 125140891
>> >> >>> > WARNING: An illegal reflective access operation has occurred
>> >> >>> > WARNING: Illegal reflective access by
>> org.apache.hadoop.security.authentication.util.KerberosUtil
>> (file:/home/rog/other/avro-tools-1.9.1.jar) to method
>> sun.security.krb5.Config.getInstance()
>> >> >>> > WARNING: Please consider reporting this to the maintainers of
>> org.apache.hadoop.security.authentication.util.KerberosUtil
>> >> >>> > WARNING: Use --illegal-access=warn to enable warnings of further
>> illegal reflective access operations
>> >> >>> > WARNING: All illegal access operations will be denied in a
>> future release
>> >> >>> > 20/01/16 11:00:37 WARN util.NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes where
>> applicable
>> >> >>> > %
>> >> >>> >
>> >> >>> > I've just verified that it's not a problem with the java
>> executable itself (I ran a program that printed to System.err and the text
>> correctly goes to the standard error).
>> >> >>> >
>> >> >>> > > Regarding the documentation, the CLI itself contains info on
>> all the available commands. Also, there are excellent online resources:
>> https://www.michael-noll.com/blog/2013/03/17/reading-and-writing-avro-files-from-the-command-line/
>> Is there anything specific that you're missing?
>> >> >>> >
>> >> >>> > There's the single line summary produced for each command by
>> running "avro-tools" with no arguments, but that's not as much info as I'd
>> ideally like. For example, it often doesn't say what file format is being
>> written or read. For some commands, the purpose is not very clear.
>> >> >>> >
>> >> >>> > For example the description of the recodec command is "Alters
>> the codec of a data file". It doesn't describe how it alters it or how one
>> might configure the alteration parameters. I managed to get some usage help
>> by passing it more than two parameters (specifying "--help" gives an
>> exception), but that doesn't provide much more info:
>> >> >>> >
>> >> >>> > % avro-tools recodec a b c
>> >> >>> > Expected at most an input file and output file.
>> >> >>> > Option             Description
>> >> >>> > ------             -----------
>> >> >>> > --codec <String>   Compression codec (default: null)
>> >> >>> > --level <Integer>  Compression level (only applies to deflate
>> and xz) (default:
>> >> >>> >                      -1)
>> >> >>> >
>> >> >>> > For the record, I'm wondering it might be possible to get
>> avrotool to tell me if one schema is compatible with another so that I can
>> check hypotheses about schema-checking in practice without having to write
>> Java code.
>> >> >>> >
>> >> >>> >   cheers,
>> >> >>> >     rog.
>> >> >>> >
>> >> >>> >
>> >> >>> > On Thu, 16 Jan 2020 at 10:30, Driesprong, Fokko
>> <fo...@driesprong.frl> wrote:
>> >> >>> >>
>> >> >>> >> Hi Rog,
>> >> >>> >>
>> >> >>> >> This is actually a warning produced by the Hadoop library, that
>> we're using. Please note that htis isn't part of the stdout:
>> >> >>> >>
>> >> >>> >> $ find /tmp/tmp
>> >> >>> >> /tmp/tmp
>> >> >>> >> /tmp/tmp/._SUCCESS.crc
>> >> >>> >>
>> /tmp/tmp/part-00000-9300fba6-ccdd-4ecc-97cb-0c3ae3631be5-c000.avro
>> >> >>> >>
>> /tmp/tmp/.part-00000-9300fba6-ccdd-4ecc-97cb-0c3ae3631be5-c000.avro.crc
>> >> >>> >> /tmp/tmp/_SUCCESS
>> >> >>> >>
>> >> >>> >> $ avro-tools tojson
>> /tmp/tmp/part-00000-9300fba6-ccdd-4ecc-97cb-0c3ae3631be5-c000.avro
>> >> >>> >> 20/01/16 11:26:10 WARN util.NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes where
>> applicable
>> >> >>> >> {"line_of_text":{"string":"Hello"}}
>> >> >>> >> {"line_of_text":{"string":"World"}}
>> >> >>> >>
>> >> >>> >> $ avro-tools tojson
>> /tmp/tmp/part-00000-9300fba6-ccdd-4ecc-97cb-0c3ae3631be5-c000.avro >
>> /tmp/tmp/data.json
>> >> >>> >> 20/01/16 11:26:20 WARN util.NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes where
>> applicable
>> >> >>> >>
>> >> >>> >> $ cat /tmp/tmp/data.json
>> >> >>> >> {"line_of_text":{"string":"Hello"}}
>> >> >>> >> {"line_of_text":{"string":"World"}}
>> >> >>> >>
>> >> >>> >> So when you pipe the data, it doesn't include the warnings.
>> >> >>> >>
>> >> >>> >> Regarding the documentation, the CLI itself contains info on
>> all the available commands. Also, there are excellent online resources:
>> https://www.michael-noll.com/blog/2013/03/17/reading-and-writing-avro-files-from-the-command-line/
>> Is there anything specific that you're missing?
>> >> >>> >>
>> >> >>> >> Hope this helps.
>> >> >>> >>
>> >> >>> >> Cheers, Fokko
>> >> >>> >>
>> >> >>> >> Op do 16 jan. 2020 om 09:30 schreef roger peppe <
>> rogpe...@gmail.com>:
>> >> >>> >>>
>> >> >>> >>> Hi,
>> >> >>> >>>
>> >> >>> >>> I've been trying to use avro-tools to verify Avro
>> implementations, and I've come across an issue. Perhaps someone here might
>> be able to help?
>> >> >>> >>>
>> >> >>> >>> When I run avro-tools with some subcommands, it prints a bunch
>> of warnings (see below) to the standard output. Does anyone know a way to
>> disable this? I'm using openjdk 11.0.5 under Ubuntu 18.04 and avro-tools
>> 1.9.1.
>> >> >>> >>>
>> >> >>> >>> The warnings are somewhat annoying because they can corrupt
>> output of tools that print to the standard output, such as recodec.
>> >> >>> >>>
>> >> >>> >>> Aside: is there any documentation for the commands in
>> avro-tools? Some seem to have some command-line help (though unfortunately
>> there doesn't seem to be a standard way of showing it), but often that help
>> often doesn't describe what the command actually does.
>> >> >>> >>>
>> >> >>> >>> Here's the output that I see:
>> >> >>> >>>
>> >> >>> >>> WARNING: An illegal reflective access operation has occurred
>> >> >>> >>> WARNING: Illegal reflective access by
>> org.apache.hadoop.security.authentication.util.KerberosUtil
>> (file:/home/rog/other/avro-tools-1.9.1.jar) to method
>> sun.security.krb5.Config.getInstance()
>> >> >>> >>> WARNING: Please consider reporting this to the maintainers of
>> org.apache.hadoop.security.authentication.util.KerberosUtil
>> >> >>> >>> WARNING: Use --illegal-access=warn to enable warnings of
>> further illegal reflective access operations
>> >> >>> >>> WARNING: All illegal access operations will be denied in a
>> future release
>> >> >>> >>> 20/01/16 08:12:39 WARN util.NativeCodeLoader: Unable to load
>> native-hadoop library for your platform... using builtin-java classes where
>> applicable
>> >> >>> >>>
>> >> >>> >>>   cheers,
>> >> >>> >>>     rog.
>> >> >>> >>>
>>
>

Reply via email to