-
- Downloads
[SPARK-15390] fix broadcast with 100 millions rows
## What changes were proposed in this pull request? When broadcast a table with more than 100 millions rows (should not ideally), the size of needed memory will overflow. This PR fix the overflow by converting it to Long when calculating the size of memory. Also add more checking in broadcast to show reasonable messages. ## How was this patch tested? Add test. Author: Davies Liu <davies@databricks.com> Closes #13182 from davies/fix_broadcast.
Showing
- sql/core/src/main/scala/org/apache/spark/sql/execution/exchange/BroadcastExchangeExec.scala 11 additions, 2 deletions.../spark/sql/execution/exchange/BroadcastExchangeExec.scala
- sql/core/src/main/scala/org/apache/spark/sql/execution/joins/HashedRelation.scala 3 additions, 2 deletions...org/apache/spark/sql/execution/joins/HashedRelation.scala
- sql/core/src/test/scala/org/apache/spark/sql/execution/joins/HashedRelationSuite.scala 15 additions, 0 deletions...pache/spark/sql/execution/joins/HashedRelationSuite.scala
Loading
Please register or sign in to comment