Ok, so we had a look with Tham at the current pyarrow and parquet-cpp
configuration objects. There is no Hadoop-like free map (this is good, I
guess). Instead, the property keys are pre-defined in most objects.
But some objects (such as HdfsConnectionConfig ,
https://github.com/apache/arrow/blob/
Hi Chris, some additional thoughts to what Antoine said.
Neal
On Wed, Jul 8, 2020 at 10:56 AM Antoine Pitrou wrote:
>
> Hi Chris,
>
> Le 08/07/2020 à 12:01, Chris Channing a écrit :
> >
> > I've looked at the contribution guidelines, but rather than arbitrarily
> > picking a jira I was hoping t
Hi,
parquet_writer.write_table(table)
This line writes a single file.
The documentation says:
This creates a single Parquet file. In practice, a Parquet dataset may
consist of many files in many directories. We can read a single file back
with read_table:
Is there a way for PyArrow to create a p
Hi Chris,
Le 08/07/2020 à 12:01, Chris Channing a écrit :
>
> I've looked at the contribution guidelines, but rather than arbitrarily
> picking a jira I was hoping that there was a more structured approach for
> newbies documented that I might have missed. A few questions that I have
> are:
As
Attendees:
* Projjal Chanda
* Rémi Dettai
* Supun Kamburugamuve
* Ben Kietzman
* Steve Kim
* Uwe Korn
* Micah Kornfield
* Rok Mihevc
* Patrick Pai
* Niranda Perera
* Neal Richardson
* Joris Van den Bossche
Discussion:
* 1.0 issues
* Custom build of pyarrow questions: directed to mailing list for d
Happy with the current version. I think this gives enough input for the board.
We have so much things happening that are much better presented in the process
of the 1.0 release.
On Wed, Jul 8, 2020, at 12:52 AM, Micah Kornfield wrote:
> Worth mentioning the website work?
>
> On Tue, Jul 7, 2020
Hi folks,
This issue is probably the one true "blocker" for the 1.0.0 release.
Ideally, all libraries should emit V5 MetadataVersion by default. How
V4 handled depends on the willingness to implement compatibility code:
* Since V4 is backwards compatible with V5 (except for unions),
libraries can
Arrow Build Report for Job nightly-2020-07-08-0
All tasks:
https://github.com/ursa-labs/crossbow/branches/all?query=nightly-2020-07-08-0
Failed Tasks:
- debian-buster-arm64:
URL:
https://github.com/ursa-labs/crossbow/branches/all?query=nightly-2020-07-08-0-travis-debian-buster-arm64
- test-
Hey Folks,
I'd like to get involved with the project from a development perspective
but I'm looking for some guidance on where/how best to start.
I've looked at the contribution guidelines, but rather than arbitrarily
picking a jira I was hoping that there was a more structured approach for
newbi