It’s one thing to get the names/values of the configurations, via the 
Spark.sql(“set -v”), but another thing to understand what each achieves and 
when and why you’ll want to use it. 

A webpage with a table and description of each is huge benefit. 

Cheers 
Jules 

Sent from my iPhone
Pardon the dumb thumb typos :)

> On Jan 16, 2020, at 11:04 AM, Shixiong(Ryan) Zhu <shixi...@databricks.com> 
> wrote:
> 
> 
> "spark.sql("set -v")" returns a Dataset that has all non-internal SQL 
> configurations. Should be pretty easy to automatically generate a SQL 
> configuration page.
> Best Regards,
> 
> Ryan
> 
> 
>> On Wed, Jan 15, 2020 at 5:47 AM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>> I think automatically creating a configuration page isn't a bad idea because 
>> I think we deprecate and remove configurations which are not created via 
>> .internal() in SQLConf anyway.
>> 
>> I already tried this automatic generation from the codes at SQL built-in 
>> functions and I'm pretty sure we can do the similar thing for configurations 
>> as well.
>> 
>> We could perhaps mimic what hadoop does 
>> https://hadoop.apache.org/docs/r2.8.0/hadoop-project-dist/hadoop-common/core-default.xml
>> 
>>> On Wed, 15 Jan 2020, 10:46 Sean Owen, <sro...@gmail.com> wrote:
>>> Some of it is intentionally undocumented, as far as I know, as an
>>> experimental option that may change, or legacy, or safety valve flag.
>>> Certainly anything that's marked an internal conf. (That does raise
>>> the question of who it's for, if you have to read source to find it.)
>>> 
>>> I don't know if we need to overhaul the conf system, but there may
>>> indeed be some confs that could legitimately be documented. I don't
>>> know which.
>>> 
>>> On Tue, Jan 14, 2020 at 7:32 PM Nicholas Chammas
>>> <nicholas.cham...@gmail.com> wrote:
>>> >
>>> > I filed SPARK-30510 thinking that we had forgotten to document an option, 
>>> > but it turns out that there's a whole bunch of stuff under SQLConf.scala 
>>> > that has no public documentation under http://spark.apache.org/docs.
>>> >
>>> > Would it be appropriate to somehow automatically generate a documentation 
>>> > page from SQLConf.scala, as Hyukjin suggested on that ticket?
>>> >
>>> > Another thought that comes to mind is moving the config definitions out 
>>> > of Scala and into a data format like YAML or JSON, and then sourcing that 
>>> > both for SQLConf as well as for whatever documentation page we want to 
>>> > generate. What do you think of that idea?
>>> >
>>> > Nick
>>> >
>>> 
>>> ---------------------------------------------------------------------
>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>> 

Reply via email to