Skip to content
Snippets Groups Projects
  • hyukjinkwon's avatar
    60472dbf
    [SPARK-21485][SQL][DOCS] Spark SQL documentation generation for built-in functions · 60472dbf
    hyukjinkwon authored
    ## What changes were proposed in this pull request?
    
    This generates a documentation for Spark SQL built-in functions.
    
    One drawback is, this requires a proper build to generate built-in function list.
    Once it is built, it only takes few seconds by `sql/create-docs.sh`.
    
    Please see https://spark-test.github.io/sparksqldoc/ that I hosted to show the output documentation.
    
    There are few more works to be done in order to make the documentation pretty, for example, separating `Arguments:` and `Examples:` but I guess this should be done within `ExpressionDescription` and `ExpressionInfo` rather than manually parsing it. I will fix these in a follow up.
    
    This requires `pip install mkdocs` to generate HTMLs from markdown files.
    
    ## How was this patch tested?
    
    Manually tested:
    
    ```
    cd docs
    jekyll build
    ```
    ,
    
    ```
    cd docs
    jekyll serve
    ```
    
    and
    
    ```
    cd sql
    create-docs.sh
    ```
    
    Author: hyukjinkwon <gurwls223@gmail.com>
    
    Closes #18702 from HyukjinKwon/SPARK-21485.
    60472dbf
    History
    [SPARK-21485][SQL][DOCS] Spark SQL documentation generation for built-in functions
    hyukjinkwon authored
    ## What changes were proposed in this pull request?
    
    This generates a documentation for Spark SQL built-in functions.
    
    One drawback is, this requires a proper build to generate built-in function list.
    Once it is built, it only takes few seconds by `sql/create-docs.sh`.
    
    Please see https://spark-test.github.io/sparksqldoc/ that I hosted to show the output documentation.
    
    There are few more works to be done in order to make the documentation pretty, for example, separating `Arguments:` and `Examples:` but I guess this should be done within `ExpressionDescription` and `ExpressionInfo` rather than manually parsing it. I will fix these in a follow up.
    
    This requires `pip install mkdocs` to generate HTMLs from markdown files.
    
    ## How was this patch tested?
    
    Manually tested:
    
    ```
    cd docs
    jekyll build
    ```
    ,
    
    ```
    cd docs
    jekyll serve
    ```
    
    and
    
    ```
    cd sql
    create-docs.sh
    ```
    
    Author: hyukjinkwon <gurwls223@gmail.com>
    
    Closes #18702 from HyukjinKwon/SPARK-21485.
.gitignore 1.25 KiB
*#*#
*.#*
*.iml
*.ipr
*.iws
*.pyc
*.pyo
*.swp
*~
.DS_Store
.cache
.classpath
.ensime
.ensime_cache/
.ensime_lucene
.generated-mima*
.idea/
.idea_modules/
.project
.pydevproject
.scala_dependencies
.settings
/lib/
R-unit-tests.log
R/unit-tests.out
R/cran-check.out
R/pkg/vignettes/sparkr-vignettes.html
R/pkg/tests/fulltests/Rplots.pdf
build/*.jar
build/apache-maven*
build/scala*
build/zinc*
cache
checkpoint
conf/*.cmd
conf/*.conf
conf/*.properties
conf/*.sh
conf/*.xml
conf/java-opts
conf/slaves
dependency-reduced-pom.xml
derby.log
dev/create-release/*final
dev/create-release/*txt
dev/pr-deps/
dist/
docs/_site
docs/api
sql/docs
sql/site
lib_managed/
lint-r-report.log
log/
logs/
out/
project/boot/
project/build/target/
project/plugins/lib_managed/
project/plugins/project/build.properties
project/plugins/src_managed/
project/plugins/target/
python/lib/pyspark.zip
python/deps
python/pyspark/python
reports/
scalastyle-on-compile.generated.xml
scalastyle-output.xml
scalastyle.txt
spark-*-bin-*.tgz
spark-tests.log
src_managed/
streaming-tests.log
target/
unit-tests.log
work/

# For Hive
TempStatsStore/
metastore/
metastore_db/
sql/hive-thriftserver/test_warehouses
warehouse/
spark-warehouse/

# For R session data
.RData
.RHistory
.Rhistory
*.Rproj
*.Rproj.*

.Rproj.user