Hi Tim,
The avro support in C++ has been on my backlog for a while.  I'm going to
try to take the first few steps towards this over the next couple of days.
Let me know if you want to collaborate on it.  C++ is a lot nicer now then
it was 8 years ago :)

Cheers,
Micah

On Tue, Jun 11, 2019 at 6:40 PM Tim Swast <sw...@google.com.invalid> wrote:

> Thanks for the advice, Wes.
>
> Unfortunately, I am about 8 years out of practice for writing any C++
> (which was part of the appeal of numba to me). Sounds like I should refresh
> my skills. I like the idea of write one, have good performance everywhere.
>
> On Tue, Jun 11, 2019 at 3:40 PM Wes McKinney <wesmck...@gmail.com> wrote:
>
> > Hi Tim,
> >
> > I'd ideally like to see the work done in the Arrow C++ library so that it
> > can be utilized by all the C++ "binders" (Python, R, C, Ruby, MATLAB).
> This
> > also means a larger labor pool of individuals to help improve and
> maintain
> > the software. There was a stalled PR around this a time back (check out
> the
> > Arrow Closed PR queue) that got stuck on some limitations in avro-c. It
> > might be more expedient to fork parts of Apache Avro and do all the
> > development inside a single codebase.
> >
> > There's a lot of folks that can provide feedback should you choose to go
> > down this route.
> >
> > Thanks
> > Wes
> >
> > On Tue, Jun 11, 2019, 4:53 PM Tim Swast <sw...@google.com.invalid>
> wrote:
> >
> > > Hi Arrow and Avro devs,
> > >
> > > I've been investigating some performance issues with the BigQuery
> Storage
> > > API (https://github.com/googleapis/google-cloud-python/issues/7805),
> and
> > > have identified that the vast majority of time is spent decoding Avro
> > into
> > > pandas dataframes.
> > > <https://github.com/googleapis/google-cloud-python/issues/7805>
> > > I've done some initial experiments by hand written parsers (inspired by
> > > https://techblog.rtbhouse.com/2017/04/18/fast-avro/) and have seen a
> > > dramatic improvement in time spent parsing.
> > >
> > > I'm considering releasing this as a separate package for the following
> > > reasons:
> > >
> > >    - Code generation + Numba is a bit of an unproven technique for
> > parsers,
> > >    so I'd like to treat this as an experiment rather than "the" package
> > to
> > > use
> > >    to parse Avro from Python.
> > >    - I don't need to handle the full Avro spec for this experiment.
> > >    Importantly, BQ Storage API only uses a schemaless reader (since the
> > > schema
> > >    is output only once, and omitted for subsequent protobuf messages)
> and
> > >    doesn't use any compression.
> > >
> > > That said, I'm open to contributing this to either pyarrow or avro if
> > > there's interest.
> > >
> > > If the answer is "no" (as I suspect it is) and I don't contribute it
> now,
> > > the package will be clearly identified as a fork of the Apache Avro
> > project
> > > and licensed Apache 2.0, so it should be easy to pull in once the
> > > techniques are proven.
> > >
> > > *  •  **Tim Swast*
> > > *  •  *Software Friendliness Engineer
> > > *  •  *Google Cloud Developer Relations
> > > *  •  *Seattle, WA, USA
> > >
> >
> --
> *  •  **Tim Swast*
> *  •  *Software Friendliness Engineer
> *  •  *Google Cloud Developer Relations
> *  •  *Seattle, WA, USA
>

Reply via email to