Well, I for one would be very interested in this project and would be happy
to contribute.


On Wed, Jan 20, 2016 at 12:09 PM Jean-Baptiste Onofré <j...@nanthrax.net>
wrote:

> Hi Sean,
>
> It's a fair point, but not present in most of the proposals. It's
> something that we can address in the "Community" section.
>
> Regards
> JB
>
> On 01/20/2016 05:55 PM, Sean Busbey wrote:
> > Great proposal. I like that your proposal includes a well presented
> > roadmap, but I don't see any goals that directly address building a
> larger
> > community. Y'all have any ideas around outreach that will help with
> > adoption?
> >
> > As a start, I recommend y'all add a section to the proposal on the wiki
> > page for "Additional Interested Contributors" so that folks who want to
> > sign up to participate in the project can do so without requesting
> > additions to the initial committer list.
> >
> > On Wed, Jan 20, 2016 at 10:32 AM, James Malone <
> > jamesmal...@google.com.invalid> wrote:
> >
> >> Hello everyone,
> >>
> >> Attached to this message is a proposed new project - Apache Dataflow, a
> >> unified programming model for data processing and integration.
> >>
> >> The text of the proposal is included below. Additionally, the proposal
> is
> >> in draft form on the wiki where we will make any required changes:
> >>
> >> https://wiki.apache.org/incubator/DataflowProposal
> >>
> >> We look forward to your feedback and input.
> >>
> >> Best,
> >>
> >> James
> >>
> >> ----
> >>
> >> = Apache Dataflow =
> >>
> >> == Abstract ==
> >>
> >> Dataflow is an open source, unified model and set of language-specific
> SDKs
> >> for defining and executing data processing workflows, and also data
> >> ingestion and integration flows, supporting Enterprise Integration
> Patterns
> >> (EIPs) and Domain Specific Languages (DSLs). Dataflow pipelines simplify
> >> the mechanics of large-scale batch and streaming data processing and can
> >> run on a number of runtimes like Apache Flink, Apache Spark, and Google
> >> Cloud Dataflow (a cloud service). Dataflow also brings DSL in different
> >> languages, allowing users to easily implement their data integration
> >> processes.
> >>
> >> == Proposal ==
> >>
> >> Dataflow is a simple, flexible, and powerful system for distributed data
> >> processing at any scale. Dataflow provides a unified programming model,
> a
> >> software development kit to define and construct data processing
> pipelines,
> >> and runners to execute Dataflow pipelines in several runtime engines,
> like
> >> Apache Spark, Apache Flink, or Google Cloud Dataflow. Dataflow can be
> used
> >> for a variety of streaming or batch data processing goals including ETL,
> >> stream analysis, and aggregate computation. The underlying programming
> >> model for Dataflow provides MapReduce-like parallelism, combined with
> >> support for powerful data windowing, and fine-grained correctness
> control.
> >>
> >> == Background ==
> >>
> >> Dataflow started as a set of Google projects focused on making data
> >> processing easier, faster, and less costly. The Dataflow model is a
> >> successor to MapReduce, FlumeJava, and Millwheel inside Google and is
> >> focused on providing a unified solution for batch and stream processing.
> >> These projects on which Dataflow is based have been published in several
> >> papers made available to the public:
> >>
> >> * MapReduce - http://research.google.com/archive/mapreduce.html
> >>
> >> * Dataflow model  - http://www.vldb.org/pvldb/vol8/p1792-Akidau.pdf
> >>
> >> * FlumeJava - http://notes.stephenholiday.com/FlumeJava.pdf
> >>
> >> * MillWheel - http://research.google.com/pubs/pub41378.html
> >>
> >> Dataflow was designed from the start to provide a portable programming
> >> layer. When you define a data processing pipeline with the Dataflow
> model,
> >> you are creating a job which is capable of being processed by any
> number of
> >> Dataflow processing engines. Several engines have been developed to run
> >> Dataflow pipelines in other open source runtimes, including a Dataflow
> >> runner for Apache Flink and Apache Spark. There is also a “direct
> runner”,
> >> for execution on the developer machine (mainly for dev/debug purposes).
> >> Another runner allows a Dataflow program to run on a managed service,
> >> Google Cloud Dataflow, in Google Cloud Platform. The Dataflow Java SDK
> is
> >> already available on GitHub, and independent from the Google Cloud
> Dataflow
> >> service. Another Python SDK is currently in active development.
> >>
> >> In this proposal, the Dataflow SDKs, model, and a set of runners will be
> >> submitted as an OSS project under the ASF. The runners which are a part
> of
> >> this proposal include those for Spark (from Cloudera), Flink (from data
> >> Artisans), and local development (from Google); the Google Cloud
> Dataflow
> >> service runner is not included in this proposal. Further references to
> >> Dataflow will refer to the Dataflow model, SDKs, and runners which are a
> >> part of this proposal (Apache Dataflow) only. The initial submission
> will
> >> contain the already-released Java SDK; Google intends to submit the
> Python
> >> SDK later in the incubation process. The Google Cloud Dataflow service
> will
> >> continue to be one of many runners for Dataflow, built on Google Cloud
> >> Platform, to run Dataflow pipelines. Necessarily, Cloud Dataflow will
> >> develop against the Apache project additions, updates, and changes.
> Google
> >> Cloud Dataflow will become one user of Apache Dataflow and will
> participate
> >> in the project openly and publicly.
> >>
> >> The Dataflow programming model has been designed with simplicity,
> >> scalability, and speed as key tenants. In the Dataflow model, you only
> need
> >> to think about four top-level concepts when constructing your data
> >> processing job:
> >>
> >> * Pipelines - The data processing job made of a series of computations
> >> including input, processing, and output
> >>
> >> * PCollections - Bounded (or unbounded) datasets which represent the
> input,
> >> intermediate and output data in pipelines
> >>
> >> * PTransforms - A data processing step in a pipeline in which one or
> more
> >> PCollections are an input and output
> >>
> >> * I/O Sources and Sinks - APIs for reading and writing data which are
> the
> >> roots and endpoints of the pipeline
> >>
> >> == Rationale ==
> >>
> >> With Dataflow, Google intended to develop a framework which allowed
> >> developers to be maximally productive in defining the processing, and
> then
> >> be able to execute the program at various levels of
> >> latency/cost/completeness without re-architecting or re-writing it. This
> >> goal was informed by Google’s past experience  developing several
> models,
> >> frameworks, and tools useful for large-scale and distributed data
> >> processing. While Google has previously published papers describing
> some of
> >> its technologies, Google decided to take a different approach with
> >> Dataflow. Google open-sourced the SDK and model alongside
> commercialization
> >> of the idea and ahead of publishing papers on the topic. As a result, a
> >> number of open source runtimes exist for Dataflow, such as the Apache
> Flink
> >> and Apache Spark runners.
> >>
> >> We believe that submitting Dataflow as an Apache project will provide an
> >> immediate, worthwhile, and substantial contribution to the open source
> >> community. As an incubating project, we believe Dataflow will have a
> better
> >> opportunity to provide a meaningful contribution to OSS and also
> integrate
> >> with other Apache projects.
> >>
> >> In the long term, we believe Dataflow can be a powerful abstraction
> layer
> >> for data processing. By providing an abstraction layer for data
> pipelines
> >> and processing, data workflows can be increasingly portable, resilient
> to
> >> breaking changes in tooling, and compatible across many execution
> engines,
> >> runtimes, and open source projects.
> >>
> >> == Initial Goals ==
> >>
> >> We are breaking our initial goals into immediate (< 2 months),
> short-term
> >> (2-4 months), and intermediate-term (> 4 months).
> >>
> >> Our immediate goals include the following:
> >>
> >> * Plan for reconciling the Dataflow Java SDK and various runners into
> one
> >> project
> >>
> >> * Plan for refactoring the existing Java SDK for better extensibility by
> >> SDK and runner writers
> >>
> >> * Validating all dependencies are ASL 2.0 or compatible
> >>
> >> * Understanding and adapting to the Apache development process
> >>
> >> Our short-term goals include:
> >>
> >> * Moving the newly-merged lists, and build utilities to Apache
> >>
> >> * Start refactoring codebase and move code to Apache Git repo
> >>
> >> * Continue development of new features, functions, and fixes in the
> >> Dataflow Java SDK, and Dataflow runners
> >>
> >> * Cleaning up the Dataflow SDK sources and crafting a roadmap and plan
> for
> >> how to include new major ideas, modules, and runtimes
> >>
> >> * Establishment of easy and clear build/test framework for Dataflow and
> >> associated runtimes; creation of testing, rollback, and validation
> policy
> >>
> >> * Analysis and design for work needed to make Dataflow a better data
> >> processing abstraction layer for multiple open source frameworks and
> >> environments
> >>
> >> Finally, we have a number of intermediate-term goals:
> >>
> >> * Roadmapping, planning, and execution of integrations with other OSS
> and
> >> non-OSS projects/products
> >>
> >> * Inclusion of additional SDK for Python, which is under active
> development
> >>
> >> == Current Status ==
> >>
> >> === Meritocracy ===
> >>
> >> Dataflow was initially developed based on ideas from many employees
> within
> >> Google. As an ASL OSS project on GitHub, the Dataflow SDK has received
> >> contributions from data Artisans, Cloudera Labs, and other individual
> >> developers. As a project under incubation, we are committed to expanding
> >> our effort to build an environment which supports a meritocracy. We are
> >> focused on engaging the community and other related projects for support
> >> and contributions. Moreover, we are committed to ensure contributors and
> >> committers to Dataflow come from a broad mix of organizations through a
> >> merit-based decision process during incubation. We believe strongly in
> the
> >> Dataflow model and are committed to growing an inclusive community of
> >> Dataflow contributors.
> >>
> >> === Community ===
> >>
> >> The core of the Dataflow Java SDK has been developed by Google for use
> with
> >> Google Cloud Dataflow. Google has active community engagement in the SDK
> >> GitHub repository (
> https://github.com/GoogleCloudPlatform/DataflowJavaSDK
> >> ),
> >> on Stack Overflow (
> >> http://stackoverflow.com/questions/tagged/google-cloud-dataflow) and
> has
> >> had contributions from a number of organizations and indivuduals.
> >>
> >> Everyday, Cloud Dataflow is actively used by a number of organizations
> and
> >> institutions for batch and stream processing of data. We believe
> acceptance
> >> will allow us to consolidate existing Dataflow-related work, grow the
> >> Dataflow community, and deepen connections between Dataflow and other
> open
> >> source projects.
> >>
> >> === Core Developers ===
> >>
> >> The core developers for Dataflow and the Dataflow runners are:
> >>
> >> * Frances Perry
> >>
> >> * Tyler Akidau
> >>
> >> * Davor Bonaci
> >>
> >> * Luke Cwik
> >>
> >> * Ben Chambers
> >>
> >> * Kenn Knowles
> >>
> >> * Dan Halperin
> >>
> >> * Daniel Mills
> >>
> >> * Mark Shields
> >>
> >> * Craig Chambers
> >>
> >> * Maximilian Michels
> >>
> >> * Tom White
> >>
> >> * Josh Wills
> >>
> >> === Alignment ===
> >>
> >> The Dataflow SDK can be used to create Dataflow pipelines which can be
> >> executed on Apache Spark or Apache Flink. Dataflow is also related to
> other
> >> Apache projects, such as Apache Crunch. We plan on expanding
> functionality
> >> for Dataflow runners, support for additional domain specific languages,
> and
> >> increased portability so Dataflow is a powerful abstraction layer for
> data
> >> processing.
> >>
> >> == Known Risks ==
> >>
> >> === Orphaned Products ===
> >>
> >> The Dataflow SDK is presently used by several organizations, from small
> >> startups to Fortune 100 companies, to construct production pipelines
> which
> >> are executed in Google Cloud Dataflow. Google has a long-term
> commitment to
> >> advance the Dataflow SDK; moreover, Dataflow is seeing increasing
> interest,
> >> development, and adoption from organizations outside of Google.
> >>
> >> === Inexperience with Open Source ===
> >>
> >> Google believes strongly in open source and the exchange of information
> to
> >> advance new ideas and work. Examples of this commitment are active OSS
> >> projects such as Chromium (https://www.chromium.org) and Kubernetes (
> >> http://kubernetes.io/). With Dataflow, we have tried to be increasingly
> >> open and forward-looking; we have published a paper in the VLDB
> conference
> >> describing the Dataflow model (
> >> http://www.vldb.org/pvldb/vol8/p1792-Akidau.pdf) and were quick to
> release
> >> the Dataflow SDK as open source software with the launch of Cloud
> Dataflow.
> >> Our submission to the Apache Software Foundation is a logical extension
> of
> >> our commitment to open source software.
> >>
> >> === Homogeneous Developers ===
> >>
> >> The majority of committers in this proposal belong to Google due to the
> >> fact that Dataflow has emerged from several internal Google projects.
> This
> >> proposal also includes committers outside of Google who are actively
> >> involved with other Apache projects, such as Hadoop, Flink, and Spark.
> We
> >> expect our entry into incubation will allow us to expand the number of
> >> individuals and organizations participating in Dataflow development.
> >> Additionally, separation of the Dataflow SDK from Google Cloud Dataflow
> >> allows us to focus on the open source SDK and model and do what is best
> for
> >> this project.
> >>
> >> === Reliance on Salaried Developers ===
> >>
> >> The Dataflow SDK and Dataflow runners have been developed primarily by
> >> salaried developers supporting the Google Cloud Dataflow project. While
> the
> >> Dataflow SDK and Cloud Dataflow have been developed by different teams
> (and
> >> this proposal would reinforce that separation) we expect our initial
> set of
> >> developers will still primarily be salaried. Contribution has not been
> >> exclusively from salaried developers, however. For example, the contrib
> >> directory of the Dataflow SDK (
> >>
> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/tree/master/contrib
> >> )
> >> contains items from free-time contributors. Moreover, seperate projects,
> >> such as ScalaFlow (https://github.com/darkjh/scalaflow) have been
> created
> >> around the Dataflow model and SDK. We expect our reliance on salaried
> >> developers will decrease over time during incubation.
> >>
> >> === Relationship with other Apache products ===
> >>
> >> Dataflow directly interoperates with or utilizes several existing Apache
> >> projects.
> >>
> >> * Build
> >>
> >> ** Apache Maven
> >>
> >> * Data I/O, Libraries
> >>
> >> ** Apache Avro
> >>
> >> ** Apache Commons
> >>
> >> * Dataflow runners
> >>
> >> ** Apache Flink
> >>
> >> ** Apache Spark
> >>
> >> Dataflow when used in batch mode shares similarities with Apache Crunch;
> >> however, Dataflow is focused on a model, SDK, and abstraction layer
> beyond
> >> Spark and Hadoop (MapReduce.) One key goal of Dataflow is to provide an
> >> intermediate abstraction layer which can easily be implemented and
> utilized
> >> across several different processing frameworks.
> >>
> >> === An excessive fascination with the Apache brand ===
> >>
> >> With this proposal we are not seeking attention or publicity. Rather, we
> >> firmly believe in the Dataflow model, SDK, and the ability to make
> Dataflow
> >> a powerful yet simple framework for data processing. While the Dataflow
> SDK
> >> and model have been open source, we believe putting code on GitHub can
> only
> >> go so far. We see the Apache community, processes, and mission as
> critical
> >> for ensuring the Dataflow SDK and model are truly community-driven,
> >> positively impactful, and innovative open source software. While Google
> has
> >> taken a number of steps to advance its various open source projects, we
> >> believe Dataflow is a great fit for the Apache Software Foundation due
> to
> >> its focus on data processing and its relationships to existing ASF
> >> projects.
> >>
> >> == Documentation ==
> >>
> >> The following documentation is relevant to this proposal. Relevant
> portion
> >> of the documentation will be contributed to the Apache Dataflow project.
> >>
> >> * Dataflow website: https://cloud.google.com/dataflow
> >>
> >> * Dataflow programming model:
> >> https://cloud.google.com/dataflow/model/programming-model
> >>
> >> * Codebases
> >>
> >> ** Dataflow Java SDK:
> >> https://github.com/GoogleCloudPlatform/DataflowJavaSDK
> >>
> >> ** Flink Dataflow runner:
> https://github.com/dataArtisans/flink-dataflow
> >>
> >> ** Spark Dataflow runner: https://github.com/cloudera/spark-dataflow
> >>
> >> * Dataflow Java SDK issue tracker:
> >> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/issues
> >>
> >> * google-cloud-dataflow tag on Stack Overflow:
> >> http://stackoverflow.com/questions/tagged/google-cloud-dataflow
> >>
> >> == Initial Source ==
> >>
> >> The initial source for Dataflow which we will submit to the Apache
> >> Foundation will include several related projects which are currently
> hosted
> >> on the GitHub repositories:
> >>
> >> * Dataflow Java SDK (
> >> https://github.com/GoogleCloudPlatform/DataflowJavaSDK)
> >>
> >> * Flink Dataflow runner (https://github.com/dataArtisans/flink-dataflow
> )
> >>
> >> * Spark Dataflow runner (https://github.com/cloudera/spark-dataflow)
> >>
> >> These projects have always been Apache 2.0 licensed. We intend to bundle
> >> all of these repositories since they are all complimentary and should be
> >> maintained in one project. Prior to our submission, we will combine all
> of
> >> these projects into a new git repository.
> >>
> >> == Source and Intellectual Property Submission Plan ==
> >>
> >> The source for the Dataflow SDK and the three runners (Spark, Flink,
> Google
> >> Cloud Dataflow) are already licensed under an Apache 2 license.
> >>
> >> * Dataflow SDK -
> >>
> https://github.com/GoogleCloudPlatform/DataflowJavaSDK/blob/master/LICENSE
> >>
> >> * Flink runner -
> >> https://github.com/dataArtisans/flink-dataflow/blob/master/LICENSE
> >>
> >> * Spark runner -
> >> https://github.com/cloudera/spark-dataflow/blob/master/LICENSE
> >>
> >> Contributors to the Dataflow SDK have also signed the Google Individual
> >> Contributor License Agreement (
> >> https://cla.developers.google.com/about/google-individual) in order to
> >> contribute to the project.
> >>
> >> With respect to trademark rights, Google does not hold a trademark on
> the
> >> phrase “Dataflow.” Based on feedback and guidance we receive during the
> >> incubation process, we are open to renaming the project if necessary for
> >> trademark or other concerns.
> >>
> >> == External Dependencies ==
> >>
> >> All external dependencies are licensed under an Apache 2.0 or
> >> Apache-compatible license. As we grow the Dataflow community we will
> >> configure our build process to require and validate all contributions
> and
> >> dependencies are licensed under the Apache 2.0 license or are under an
> >> Apache-compatible license.
> >>
> >> == Required Resources ==
> >>
> >> === Mailing Lists ===
> >>
> >> We currently use a mix of mailing lists. We will migrate our existing
> >> mailing lists to the following:
> >>
> >> * d...@dataflow.incubator.apache.org
> >>
> >> * u...@dataflow.incubator.apache.org
> >>
> >> * priv...@dataflow.incubator.apache.org
> >>
> >> * comm...@dataflow.incubator.apache.org
> >>
> >> === Source Control ===
> >>
> >> The Dataflow team currently uses Git and would like to continue to do
> so.
> >> We request a Git repository for Dataflow with mirroring to GitHub
> enabled.
> >>
> >> === Issue Tracking ===
> >>
> >> We request the creation of an Apache-hosted JIRA. The Dataflow project
> is
> >> currently using both a public GitHub issue tracker and internal Google
> >> issue tracking. We will migrate and combine from these two sources to
> the
> >> Apache JIRA.
> >>
> >> == Initial Committers ==
> >>
> >> * Aljoscha Krettek     [aljos...@apache.org]
> >>
> >> * Amit Sela            [amitsel...@gmail.com]
> >>
> >> * Ben Chambers         [bchamb...@google.com]
> >>
> >> * Craig Chambers       [chamb...@google.com]
> >>
> >> * Dan Halperin         [dhalp...@google.com]
> >>
> >> * Davor Bonaci         [da...@google.com]
> >>
> >> * Frances Perry        [f...@google.com]
> >>
> >> * James Malone         [jamesmal...@google.com]
> >>
> >> * Jean-Baptiste Onofré [jbono...@apache.org]
> >>
> >> * Josh Wills           [jwi...@apache.org]
> >>
> >> * Kostas Tzoumas       [kos...@data-artisans.com]
> >>
> >> * Kenneth Knowles      [k...@google.com]
> >>
> >> * Luke Cwik            [lc...@google.com]
> >>
> >> * Maximilian Michels   [m...@apache.org]
> >>
> >> * Stephan Ewen         [step...@data-artisans.com]
> >>
> >> * Tom White            [t...@cloudera.com]
> >>
> >> * Tyler Akidau         [taki...@google.com]
> >>
> >> == Affiliations ==
> >>
> >> The initial committers are from six organizations. Google developed
> >> Dataflow and the Dataflow SDK, data Artisans developed the Flink runner,
> >> and Cloudera (Labs) developed the Spark runner.
> >>
> >> * Cloudera
> >>
> >> ** Tom White
> >>
> >> * Data Artisans
> >>
> >> ** Aljoscha Krettek
> >>
> >> ** Kostas Tzoumas
> >>
> >> ** Maximilian Michels
> >>
> >> ** Stephan Ewen
> >>
> >> * Google
> >>
> >> ** Ben Chambers
> >>
> >> ** Dan Halperin
> >>
> >> ** Davor Bonaci
> >>
> >> ** Frances Perry
> >>
> >> ** James Malone
> >>
> >> ** Kenneth Knowles
> >>
> >> ** Luke Cwik
> >>
> >> ** Tyler Akidau
> >>
> >> * PayPal
> >>
> >> ** Amit Sela
> >>
> >> * Slack
> >>
> >> ** Josh Wills
> >>
> >> * Talend
> >>
> >> ** Jean-Baptiste Onofré
> >>
> >> == Sponsors ==
> >>
> >> === Champion ===
> >>
> >> * Jean-Baptiste Onofre      [jbono...@apache.org]
> >>
> >> === Nominated Mentors ===
> >>
> >> * Jim Jagielski           [j...@apache.org]
> >>
> >> * Venkatesh Seetharam     [venkat...@apache.org]
> >>
> >> * Bertrand Delacretaz     [bdelacre...@apache.org]
> >>
> >> * Ted Dunning             [tdunn...@apache.org]
> >>
> >> === Sponsoring Entity ===
> >>
> >> The Apache Incubator
> >>
> >
> >
> >
>
> --
> Jean-Baptiste Onofré
> jbono...@apache.org
> http://blog.nanthrax.net
> Talend - http://www.talend.com
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org
> For additional commands, e-mail: general-h...@incubator.apache.org
>
>

Reply via email to