From 39ac56fc60734d0e095314fc38a7b36fbb4c80f7 Mon Sep 17 00:00:00 2001 From: Reynold Xin <rxin@databricks.com> Date: Mon, 18 Jan 2016 17:10:32 -0800 Subject: [PATCH] [SPARK-12889][SQL] Rename ParserDialect -> ParserInterface. Based on discussions in #10801, I'm submitting a pull request to rename ParserDialect to ParserInterface. Author: Reynold Xin <rxin@databricks.com> Closes #10817 from rxin/SPARK-12889. --- .../apache/spark/sql/catalyst/AbstractSparkSQLParser.scala | 2 +- .../main/scala/org/apache/spark/sql/catalyst/CatalystQl.scala | 2 +- .../catalyst/{ParserDialect.scala => ParserInterface.scala} | 2 +- sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala | 2 +- .../scala/org/apache/spark/sql/execution/SparkSQLParser.scala | 4 ++-- .../apache/spark/sql/execution/datasources/DDLParser.scala | 4 ++-- .../main/scala/org/apache/spark/sql/hive/HiveContext.scala | 4 ++-- 7 files changed, 10 insertions(+), 10 deletions(-) rename sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/{ParserDialect.scala => ParserInterface.scala} (98%) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/AbstractSparkSQLParser.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/AbstractSparkSQLParser.scala index 9443369808..38fa5cb585 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/AbstractSparkSQLParser.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/AbstractSparkSQLParser.scala @@ -26,7 +26,7 @@ import scala.util.parsing.input.CharArrayReader.EofCh import org.apache.spark.sql.catalyst.plans.logical._ private[sql] abstract class AbstractSparkSQLParser - extends StandardTokenParsers with PackratParsers with ParserDialect { + extends StandardTokenParsers with PackratParsers with ParserInterface { def parsePlan(input: String): LogicalPlan = synchronized { // Initialize the Keywords. diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystQl.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystQl.scala index 4bc721d0b2..5fb41f7e4b 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystQl.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/CatalystQl.scala @@ -33,7 +33,7 @@ import org.apache.spark.util.random.RandomSampler /** * This class translates SQL to Catalyst [[LogicalPlan]]s or [[Expression]]s. */ -private[sql] class CatalystQl(val conf: ParserConf = SimpleParserConf()) extends ParserDialect { +private[sql] class CatalystQl(val conf: ParserConf = SimpleParserConf()) extends ParserInterface { object Token { def unapply(node: ASTNode): Some[(String, List[ASTNode])] = { CurrentOrigin.setPosition(node.line, node.positionInLine) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserDialect.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserInterface.scala similarity index 98% rename from sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserDialect.scala rename to sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserInterface.scala index 3aa141af63..24ec452c4d 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserDialect.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/ParserInterface.scala @@ -23,7 +23,7 @@ import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan /** * Interface for a parser. */ -trait ParserDialect { +trait ParserInterface { /** Creates LogicalPlan for a given SQL string. */ def parsePlan(sqlText: String): LogicalPlan diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala index b8c8c78b91..147e3557b6 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala @@ -203,7 +203,7 @@ class SQLContext private[sql]( protected[sql] lazy val optimizer: Optimizer = new SparkOptimizer(this) @transient - protected[sql] val sqlParser: ParserDialect = new SparkSQLParser(new SparkQl(conf)) + protected[sql] val sqlParser: ParserInterface = new SparkSQLParser(new SparkQl(conf)) @transient protected[sql] val ddlParser: DDLParser = new DDLParser(sqlParser) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkSQLParser.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkSQLParser.scala index 1af2c756cd..d2d8271563 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkSQLParser.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/SparkSQLParser.scala @@ -19,7 +19,7 @@ package org.apache.spark.sql.execution import scala.util.parsing.combinator.RegexParsers -import org.apache.spark.sql.catalyst.{AbstractSparkSQLParser, ParserDialect, TableIdentifier} +import org.apache.spark.sql.catalyst.{AbstractSparkSQLParser, ParserInterface, TableIdentifier} import org.apache.spark.sql.catalyst.expressions.{Attribute, AttributeReference, Expression} import org.apache.spark.sql.catalyst.plans.logical import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan @@ -33,7 +33,7 @@ import org.apache.spark.sql.types.StringType * parameter because this allows us to return a different dialect if we * have to. */ -class SparkSQLParser(fallback: => ParserDialect) extends AbstractSparkSQLParser { +class SparkSQLParser(fallback: => ParserInterface) extends AbstractSparkSQLParser { override def parseExpression(sql: String): Expression = fallback.parseExpression(sql) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DDLParser.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DDLParser.scala index 4dea947f6a..f4766b0370 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DDLParser.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DDLParser.scala @@ -22,7 +22,7 @@ import scala.util.matching.Regex import org.apache.spark.Logging import org.apache.spark.sql.SaveMode -import org.apache.spark.sql.catalyst.{AbstractSparkSQLParser, ParserDialect, TableIdentifier} +import org.apache.spark.sql.catalyst.{AbstractSparkSQLParser, ParserInterface, TableIdentifier} import org.apache.spark.sql.catalyst.analysis.UnresolvedRelation import org.apache.spark.sql.catalyst.expressions.Expression import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan @@ -32,7 +32,7 @@ import org.apache.spark.sql.types._ /** * A parser for foreign DDL commands. */ -class DDLParser(fallback: => ParserDialect) +class DDLParser(fallback: => ParserInterface) extends AbstractSparkSQLParser with DataTypeParser with Logging { override def parseExpression(sql: String): Expression = fallback.parseExpression(sql) diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala index dd536b78c7..eaca3c9269 100644 --- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala +++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala @@ -42,7 +42,7 @@ import org.apache.spark.api.java.JavaSparkContext import org.apache.spark.sql._ import org.apache.spark.sql.SQLConf.SQLConfEntry import org.apache.spark.sql.SQLConf.SQLConfEntry._ -import org.apache.spark.sql.catalyst.{InternalRow, ParserDialect} +import org.apache.spark.sql.catalyst.{InternalRow, ParserInterface} import org.apache.spark.sql.catalyst.analysis._ import org.apache.spark.sql.catalyst.expressions.{Expression, LeafExpression} import org.apache.spark.sql.catalyst.expressions.codegen.CodegenFallback @@ -546,7 +546,7 @@ class HiveContext private[hive]( } @transient - protected[sql] override val sqlParser: ParserDialect = { + protected[sql] override val sqlParser: ParserInterface = { new SparkSQLParser(new ExtendedHiveQlParser(this)) } -- GitLab