I think automatically creating a configuration page isn't a bad idea
because I think we deprecate and remove configurations which are not
created via .internal() in SQLConf anyway.

I already tried this automatic generation from the codes at SQL built-in
functions and I'm pretty sure we can do the similar thing for
configurations as well.

We could perhaps mimic what hadoop does
https://hadoop.apache.org/docs/r2.8.0/hadoop-project-dist/hadoop-common/core-default.xml

On Wed, 15 Jan 2020, 10:46 Sean Owen, <sro...@gmail.com> wrote:

> Some of it is intentionally undocumented, as far as I know, as an
> experimental option that may change, or legacy, or safety valve flag.
> Certainly anything that's marked an internal conf. (That does raise
> the question of who it's for, if you have to read source to find it.)
>
> I don't know if we need to overhaul the conf system, but there may
> indeed be some confs that could legitimately be documented. I don't
> know which.
>
> On Tue, Jan 14, 2020 at 7:32 PM Nicholas Chammas
> <nicholas.cham...@gmail.com> wrote:
> >
> > I filed SPARK-30510 thinking that we had forgotten to document an
> option, but it turns out that there's a whole bunch of stuff under
> SQLConf.scala that has no public documentation under
> http://spark.apache.org/docs.
> >
> > Would it be appropriate to somehow automatically generate a
> documentation page from SQLConf.scala, as Hyukjin suggested on that ticket?
> >
> > Another thought that comes to mind is moving the config definitions out
> of Scala and into a data format like YAML or JSON, and then sourcing that
> both for SQLConf as well as for whatever documentation page we want to
> generate. What do you think of that idea?
> >
> > Nick
> >
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to