-
- Downloads
[SPARK-18856][SQL] non-empty partitioned table should not report zero size
## What changes were proposed in this pull request? In `DataSource`, if the table is not analyzed, we will use 0 as the default value for table size. This is dangerous, we may broadcast a large table and cause OOM. We should use `defaultSizeInBytes` instead. ## How was this patch tested? new regression test Author: Wenchen Fan <wenchen@databricks.com> Closes #16280 from cloud-fan/bug. (cherry picked from commit d6f11a12) Signed-off-by:Reynold Xin <rxin@databricks.com>
Showing
- sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala 2 additions, 1 deletion...g/apache/spark/sql/execution/datasources/DataSource.scala
- sql/core/src/test/scala/org/apache/spark/sql/StatisticsCollectionSuite.scala 18 additions, 0 deletions...cala/org/apache/spark/sql/StatisticsCollectionSuite.scala
Please register or sign in to comment