diff --git a/core/src/main/scala/org/apache/spark/util/collection/AppendOnlyMap.scala b/core/src/main/scala/org/apache/spark/util/collection/AppendOnlyMap.scala
index ad38250ad339fc0027b5a26b96b5363e23cda183..1a6f1c2b55799433570d3c82d5883a97d489596e 100644
--- a/core/src/main/scala/org/apache/spark/util/collection/AppendOnlyMap.scala
+++ b/core/src/main/scala/org/apache/spark/util/collection/AppendOnlyMap.scala
@@ -40,6 +40,8 @@ class AppendOnlyMap[K, V](initialCapacity: Int = 64)
   require(initialCapacity <= (1 << 29), "Can't make capacity bigger than 2^29 elements")
   require(initialCapacity >= 1, "Invalid initial capacity")
 
+  private val LOAD_FACTOR = 0.7
+
   private var capacity = nextPowerOf2(initialCapacity)
   private var mask = capacity - 1
   private var curSize = 0
@@ -57,8 +59,6 @@ class AppendOnlyMap[K, V](initialCapacity: Int = 64)
   private var destroyed = false
   private val destructionMessage = "Map state is invalid from destructive sorting!"
 
-  private val LOAD_FACTOR = 0.7
-
   /** Get the value for a given key */
   def apply(key: K): V = {
     assert(!destroyed, destructionMessage)