diff --git a/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/HeapMemoryAllocator.java b/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/HeapMemoryAllocator.java index 3cd4264680bfc18a20ef50f35baf23e24147a6c3..355748238540b88740e419dca94a73c93b77e37d 100644 --- a/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/HeapMemoryAllocator.java +++ b/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/HeapMemoryAllocator.java @@ -24,7 +24,6 @@ import java.util.LinkedList; import java.util.Map; import org.apache.spark.unsafe.Platform; -import org.apache.spark.unsafe.memory.MemoryAllocator; /** * A simple {@link MemoryAllocator} that can allocate up to 16GB using a JVM long primitive array. diff --git a/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/MemoryAllocator.java b/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/MemoryAllocator.java index 8bd2b06db8b8b7fdab29fa14803833af26de3b82..7b588681d979089aa4c63205837659e730e93a77 100644 --- a/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/MemoryAllocator.java +++ b/common/unsafe/src/main/java/org/apache/spark/unsafe/memory/MemoryAllocator.java @@ -23,12 +23,12 @@ public interface MemoryAllocator { * Whether to fill newly allocated and deallocated memory with 0xa5 and 0x5a bytes respectively. * This helps catch misuse of uninitialized or freed memory, but imposes some overhead. */ - public static final boolean MEMORY_DEBUG_FILL_ENABLED = Boolean.parseBoolean( + boolean MEMORY_DEBUG_FILL_ENABLED = Boolean.parseBoolean( System.getProperty("spark.memory.debugFill", "false")); // Same as jemalloc's debug fill values. - public static final byte MEMORY_DEBUG_FILL_CLEAN_VALUE = (byte)0xa5; - public static final byte MEMORY_DEBUG_FILL_FREED_VALUE = (byte)0x5a; + byte MEMORY_DEBUG_FILL_CLEAN_VALUE = (byte)0xa5; + byte MEMORY_DEBUG_FILL_FREED_VALUE = (byte)0x5a; /** * Allocates a contiguous block of memory. Note that the allocated memory is not guaranteed diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala index 76d7f5cbc3db99a3a22af9ba7e78b732accfa8f8..772e031ea77d31ccce7d128b0f2c6571c82dc93c 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileFormat.scala @@ -780,8 +780,7 @@ private[sql] object ParquetFileFormat extends Logging { val assumeBinaryIsString = sparkSession.sessionState.conf.isParquetBinaryAsString val assumeInt96IsTimestamp = sparkSession.sessionState.conf.isParquetINT96AsTimestamp val writeLegacyParquetFormat = sparkSession.sessionState.conf.writeLegacyParquetFormat - val serializedConf = - new SerializableConfiguration(sparkSession.sessionState.newHadoopConf()) + val serializedConf = new SerializableConfiguration(sparkSession.sessionState.newHadoopConf()) // !! HACK ALERT !! // diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceTest.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceTest.scala index 206d03ea98e6908a369eaab2fb0268d8e22b50cc..cc77d3c4b91acdf67e0a5dc9d2c0501517f366f7 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceTest.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/DataSourceTest.scala @@ -18,7 +18,6 @@ package org.apache.spark.sql.sources import org.apache.spark.sql._ -import org.apache.spark.sql.internal.SQLConf private[sql] abstract class DataSourceTest extends QueryTest {