From e76679a814f5a0903c5f93d9a482f5ddc56fe0d2 Mon Sep 17 00:00:00 2001 From: Reynold Xin <rxin@databricks.com> Date: Tue, 15 Mar 2016 10:39:07 +0800 Subject: [PATCH] [SPARK-13880][SPARK-13881][SQL] Rename DataFrame.scala Dataset.scala, and remove LegacyFunctions ## What changes were proposed in this pull request? 1. Rename DataFrame.scala Dataset.scala, since the class is now named Dataset. 2. Remove LegacyFunctions. It was introduced in Spark 1.6 for backward compatibility, and can be removed in Spark 2.0. ## How was this patch tested? Should be covered by existing unit/integration tests. Author: Reynold Xin <rxin@databricks.com> Closes #11704 from rxin/SPARK-13880. --- project/MimaExcludes.scala | 1 + .../sql/{DataFrame.scala => Dataset.scala} | 0 .../org/apache/spark/sql/functions.scala | 23 ++----------------- 3 files changed, 3 insertions(+), 21 deletions(-) rename sql/core/src/main/scala/org/apache/spark/sql/{DataFrame.scala => Dataset.scala} (100%) diff --git a/project/MimaExcludes.scala b/project/MimaExcludes.scala index a9973bc24c..2a4a874fef 100644 --- a/project/MimaExcludes.scala +++ b/project/MimaExcludes.scala @@ -314,6 +314,7 @@ object MimaExcludes { ProblemFilters.exclude[MissingClassProblem]("org.apache.spark.sql.DataFrame"), ProblemFilters.exclude[MissingClassProblem]("org.apache.spark.sql.DataFrame$"), + ProblemFilters.exclude[MissingClassProblem]("org.apache.spark.sql.LegacyFunctions"), ProblemFilters.exclude[IncompatibleMethTypeProblem]("org.apache.spark.mllib.evaluation.MultilabelMetrics.this"), ProblemFilters.exclude[IncompatibleResultTypeProblem]("org.apache.spark.ml.classification.LogisticRegressionSummary.predictions"), diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala b/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala similarity index 100% rename from sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala rename to sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala diff --git a/sql/core/src/main/scala/org/apache/spark/sql/functions.scala b/sql/core/src/main/scala/org/apache/spark/sql/functions.scala index 737e125f6c..326c1e5a7c 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/functions.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/functions.scala @@ -33,25 +33,6 @@ import org.apache.spark.sql.expressions.UserDefinedFunction import org.apache.spark.sql.types._ import org.apache.spark.util.Utils -/** - * Ensures that java functions signatures for methods that now return a [[TypedColumn]] still have - * legacy equivalents in bytecode. This compatibility is done by forcing the compiler to generate - * "bridge" methods due to the use of covariant return types. - * - * {{{ - * // In LegacyFunctions: - * public abstract org.apache.spark.sql.Column avg(java.lang.String); - * - * // In functions: - * public static org.apache.spark.sql.TypedColumn<java.lang.Object, java.lang.Object> avg(...); - * }}} - * - * This allows us to use the same functions both in typed [[Dataset]] operations and untyped - * [[DataFrame]] operations when the return type for a given function is statically known. - */ -private[sql] abstract class LegacyFunctions { - def count(columnName: String): Column -} /** * :: Experimental :: @@ -72,7 +53,7 @@ private[sql] abstract class LegacyFunctions { */ @Experimental // scalastyle:off -object functions extends LegacyFunctions { +object functions { // scalastyle:on private def withExpr(expr: Expression): Column = Column(expr) @@ -287,7 +268,7 @@ object functions extends LegacyFunctions { * @since 1.3.0 */ def count(columnName: String): TypedColumn[Any, Long] = - count(Column(columnName)).as(ExpressionEncoder[Long]) + count(Column(columnName)).as(ExpressionEncoder[Long]()) /** * Aggregate function: returns the number of distinct items in a group. -- GitLab