diff --git a/sql/core/src/test/java/test/org/apache/spark/sql/JavaDataFrameSuite.java b/sql/core/src/test/java/test/org/apache/spark/sql/JavaDataFrameSuite.java
index 318b53cdbbaa04c6cadc250cdf37f3a8f9d7e848..c44fc3d393862a320eb143f58d811483bf2a5465 100644
--- a/sql/core/src/test/java/test/org/apache/spark/sql/JavaDataFrameSuite.java
+++ b/sql/core/src/test/java/test/org/apache/spark/sql/JavaDataFrameSuite.java
@@ -327,23 +327,23 @@ public class JavaDataFrameSuite {
 
   @Test
   public void testGenericLoad() {
-    Dataset<Row> df1 = spark.read().format("text").load(getResource("text-suite.txt"));
+    Dataset<Row> df1 = spark.read().format("text").load(getResource("test-data/text-suite.txt"));
     Assert.assertEquals(4L, df1.count());
 
     Dataset<Row> df2 = spark.read().format("text").load(
-      getResource("text-suite.txt"),
-      getResource("text-suite2.txt"));
+      getResource("test-data/text-suite.txt"),
+      getResource("test-data/text-suite2.txt"));
     Assert.assertEquals(5L, df2.count());
   }
 
   @Test
   public void testTextLoad() {
-    Dataset<String> ds1 = spark.read().textFile(getResource("text-suite.txt"));
+    Dataset<String> ds1 = spark.read().textFile(getResource("test-data/text-suite.txt"));
     Assert.assertEquals(4L, ds1.count());
 
     Dataset<String> ds2 = spark.read().textFile(
-      getResource("text-suite.txt"),
-      getResource("text-suite2.txt"));
+      getResource("test-data/text-suite.txt"),
+      getResource("test-data/text-suite2.txt"));
     Assert.assertEquals(5L, ds2.count());
   }
 
diff --git a/sql/core/src/test/resources/old-repeated.parquet b/sql/core/src/test/resources/old-repeated.parquet
deleted file mode 100644
index 213f1a90291b30a8a3161b51c38f008f3ae9f6e5..0000000000000000000000000000000000000000
Binary files a/sql/core/src/test/resources/old-repeated.parquet and /dev/null differ
diff --git a/sql/core/src/test/resources/bool.csv b/sql/core/src/test/resources/test-data/bool.csv
similarity index 100%
rename from sql/core/src/test/resources/bool.csv
rename to sql/core/src/test/resources/test-data/bool.csv
diff --git a/sql/core/src/test/resources/cars-alternative.csv b/sql/core/src/test/resources/test-data/cars-alternative.csv
similarity index 100%
rename from sql/core/src/test/resources/cars-alternative.csv
rename to sql/core/src/test/resources/test-data/cars-alternative.csv
diff --git a/sql/core/src/test/resources/cars-blank-column-name.csv b/sql/core/src/test/resources/test-data/cars-blank-column-name.csv
similarity index 100%
rename from sql/core/src/test/resources/cars-blank-column-name.csv
rename to sql/core/src/test/resources/test-data/cars-blank-column-name.csv
diff --git a/sql/core/src/test/resources/cars-malformed.csv b/sql/core/src/test/resources/test-data/cars-malformed.csv
similarity index 100%
rename from sql/core/src/test/resources/cars-malformed.csv
rename to sql/core/src/test/resources/test-data/cars-malformed.csv
diff --git a/sql/core/src/test/resources/cars-null.csv b/sql/core/src/test/resources/test-data/cars-null.csv
similarity index 100%
rename from sql/core/src/test/resources/cars-null.csv
rename to sql/core/src/test/resources/test-data/cars-null.csv
diff --git a/sql/core/src/test/resources/cars-unbalanced-quotes.csv b/sql/core/src/test/resources/test-data/cars-unbalanced-quotes.csv
similarity index 100%
rename from sql/core/src/test/resources/cars-unbalanced-quotes.csv
rename to sql/core/src/test/resources/test-data/cars-unbalanced-quotes.csv
diff --git a/sql/core/src/test/resources/cars.csv b/sql/core/src/test/resources/test-data/cars.csv
similarity index 100%
rename from sql/core/src/test/resources/cars.csv
rename to sql/core/src/test/resources/test-data/cars.csv
diff --git a/sql/core/src/test/resources/cars.tsv b/sql/core/src/test/resources/test-data/cars.tsv
similarity index 100%
rename from sql/core/src/test/resources/cars.tsv
rename to sql/core/src/test/resources/test-data/cars.tsv
diff --git a/sql/core/src/test/resources/cars_iso-8859-1.csv b/sql/core/src/test/resources/test-data/cars_iso-8859-1.csv
similarity index 100%
rename from sql/core/src/test/resources/cars_iso-8859-1.csv
rename to sql/core/src/test/resources/test-data/cars_iso-8859-1.csv
diff --git a/sql/core/src/test/resources/comments.csv b/sql/core/src/test/resources/test-data/comments.csv
similarity index 100%
rename from sql/core/src/test/resources/comments.csv
rename to sql/core/src/test/resources/test-data/comments.csv
diff --git a/sql/core/src/test/resources/dates.csv b/sql/core/src/test/resources/test-data/dates.csv
similarity index 100%
rename from sql/core/src/test/resources/dates.csv
rename to sql/core/src/test/resources/test-data/dates.csv
diff --git a/sql/core/src/test/resources/dec-in-fixed-len.parquet b/sql/core/src/test/resources/test-data/dec-in-fixed-len.parquet
similarity index 100%
rename from sql/core/src/test/resources/dec-in-fixed-len.parquet
rename to sql/core/src/test/resources/test-data/dec-in-fixed-len.parquet
diff --git a/sql/core/src/test/resources/dec-in-i32.parquet b/sql/core/src/test/resources/test-data/dec-in-i32.parquet
similarity index 100%
rename from sql/core/src/test/resources/dec-in-i32.parquet
rename to sql/core/src/test/resources/test-data/dec-in-i32.parquet
diff --git a/sql/core/src/test/resources/dec-in-i64.parquet b/sql/core/src/test/resources/test-data/dec-in-i64.parquet
similarity index 100%
rename from sql/core/src/test/resources/dec-in-i64.parquet
rename to sql/core/src/test/resources/test-data/dec-in-i64.parquet
diff --git a/sql/core/src/test/resources/decimal.csv b/sql/core/src/test/resources/test-data/decimal.csv
similarity index 100%
rename from sql/core/src/test/resources/decimal.csv
rename to sql/core/src/test/resources/test-data/decimal.csv
diff --git a/sql/core/src/test/resources/disable_comments.csv b/sql/core/src/test/resources/test-data/disable_comments.csv
similarity index 100%
rename from sql/core/src/test/resources/disable_comments.csv
rename to sql/core/src/test/resources/test-data/disable_comments.csv
diff --git a/sql/core/src/test/resources/empty.csv b/sql/core/src/test/resources/test-data/empty.csv
similarity index 100%
rename from sql/core/src/test/resources/empty.csv
rename to sql/core/src/test/resources/test-data/empty.csv
diff --git a/sql/core/src/test/resources/nested-array-struct.parquet b/sql/core/src/test/resources/test-data/nested-array-struct.parquet
similarity index 100%
rename from sql/core/src/test/resources/nested-array-struct.parquet
rename to sql/core/src/test/resources/test-data/nested-array-struct.parquet
diff --git a/sql/core/src/test/resources/numbers.csv b/sql/core/src/test/resources/test-data/numbers.csv
similarity index 100%
rename from sql/core/src/test/resources/numbers.csv
rename to sql/core/src/test/resources/test-data/numbers.csv
diff --git a/sql/core/src/test/resources/old-repeated-int.parquet b/sql/core/src/test/resources/test-data/old-repeated-int.parquet
similarity index 100%
rename from sql/core/src/test/resources/old-repeated-int.parquet
rename to sql/core/src/test/resources/test-data/old-repeated-int.parquet
diff --git a/sql/core/src/test/resources/old-repeated-message.parquet b/sql/core/src/test/resources/test-data/old-repeated-message.parquet
similarity index 100%
rename from sql/core/src/test/resources/old-repeated-message.parquet
rename to sql/core/src/test/resources/test-data/old-repeated-message.parquet
diff --git a/sql/core/src/test/resources/parquet-thrift-compat.snappy.parquet b/sql/core/src/test/resources/test-data/parquet-thrift-compat.snappy.parquet
similarity index 100%
rename from sql/core/src/test/resources/parquet-thrift-compat.snappy.parquet
rename to sql/core/src/test/resources/test-data/parquet-thrift-compat.snappy.parquet
diff --git a/sql/core/src/test/resources/proto-repeated-string.parquet b/sql/core/src/test/resources/test-data/proto-repeated-string.parquet
similarity index 100%
rename from sql/core/src/test/resources/proto-repeated-string.parquet
rename to sql/core/src/test/resources/test-data/proto-repeated-string.parquet
diff --git a/sql/core/src/test/resources/proto-repeated-struct.parquet b/sql/core/src/test/resources/test-data/proto-repeated-struct.parquet
similarity index 100%
rename from sql/core/src/test/resources/proto-repeated-struct.parquet
rename to sql/core/src/test/resources/test-data/proto-repeated-struct.parquet
diff --git a/sql/core/src/test/resources/proto-struct-with-array-many.parquet b/sql/core/src/test/resources/test-data/proto-struct-with-array-many.parquet
similarity index 100%
rename from sql/core/src/test/resources/proto-struct-with-array-many.parquet
rename to sql/core/src/test/resources/test-data/proto-struct-with-array-many.parquet
diff --git a/sql/core/src/test/resources/proto-struct-with-array.parquet b/sql/core/src/test/resources/test-data/proto-struct-with-array.parquet
similarity index 100%
rename from sql/core/src/test/resources/proto-struct-with-array.parquet
rename to sql/core/src/test/resources/test-data/proto-struct-with-array.parquet
diff --git a/sql/core/src/test/resources/simple_sparse.csv b/sql/core/src/test/resources/test-data/simple_sparse.csv
similarity index 100%
rename from sql/core/src/test/resources/simple_sparse.csv
rename to sql/core/src/test/resources/test-data/simple_sparse.csv
diff --git a/sql/core/src/test/resources/text-partitioned/year=2014/data.txt b/sql/core/src/test/resources/test-data/text-partitioned/year=2014/data.txt
similarity index 100%
rename from sql/core/src/test/resources/text-partitioned/year=2014/data.txt
rename to sql/core/src/test/resources/test-data/text-partitioned/year=2014/data.txt
diff --git a/sql/core/src/test/resources/text-partitioned/year=2015/data.txt b/sql/core/src/test/resources/test-data/text-partitioned/year=2015/data.txt
similarity index 100%
rename from sql/core/src/test/resources/text-partitioned/year=2015/data.txt
rename to sql/core/src/test/resources/test-data/text-partitioned/year=2015/data.txt
diff --git a/sql/core/src/test/resources/text-suite.txt b/sql/core/src/test/resources/test-data/text-suite.txt
similarity index 100%
rename from sql/core/src/test/resources/text-suite.txt
rename to sql/core/src/test/resources/test-data/text-suite.txt
diff --git a/sql/core/src/test/resources/text-suite2.txt b/sql/core/src/test/resources/test-data/text-suite2.txt
similarity index 100%
rename from sql/core/src/test/resources/text-suite2.txt
rename to sql/core/src/test/resources/test-data/text-suite2.txt
diff --git a/sql/core/src/test/resources/unescaped-quotes.csv b/sql/core/src/test/resources/test-data/unescaped-quotes.csv
similarity index 100%
rename from sql/core/src/test/resources/unescaped-quotes.csv
rename to sql/core/src/test/resources/test-data/unescaped-quotes.csv
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
index e14e84e0a76429366a61111021c4e708b7871c7b..ce1f7c5082ca12f8cc003355541b798341912e63 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
@@ -677,7 +677,8 @@ class DDLSuite extends QueryTest with SharedSQLContext with BeforeAndAfterEach {
   }
 
   test("create temporary view using") {
-    val csvFile = Thread.currentThread().getContextClassLoader.getResource("cars.csv").toString()
+    val csvFile =
+      Thread.currentThread().getContextClassLoader.getResource("test-data/cars.csv").toString
     withView("testview") {
       sql(s"CREATE OR REPLACE TEMPORARY VIEW testview (c1: String, c2: String)  USING " +
         "org.apache.spark.sql.execution.datasources.csv.CSVFileFormat  " +
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
index 311f1fa8d2aff7d7e5863060e58c5a1c759abf82..8cd76ddf20f045c750c88a20830815107bda4809 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVSuite.scala
@@ -33,23 +33,23 @@ import org.apache.spark.sql.types._
 class CSVSuite extends QueryTest with SharedSQLContext with SQLTestUtils {
   import testImplicits._
 
-  private val carsFile = "cars.csv"
-  private val carsMalformedFile = "cars-malformed.csv"
-  private val carsFile8859 = "cars_iso-8859-1.csv"
-  private val carsTsvFile = "cars.tsv"
-  private val carsAltFile = "cars-alternative.csv"
-  private val carsUnbalancedQuotesFile = "cars-unbalanced-quotes.csv"
-  private val carsNullFile = "cars-null.csv"
-  private val carsBlankColName = "cars-blank-column-name.csv"
-  private val emptyFile = "empty.csv"
-  private val commentsFile = "comments.csv"
-  private val disableCommentsFile = "disable_comments.csv"
-  private val boolFile = "bool.csv"
-  private val decimalFile = "decimal.csv"
-  private val simpleSparseFile = "simple_sparse.csv"
-  private val numbersFile = "numbers.csv"
-  private val datesFile = "dates.csv"
-  private val unescapedQuotesFile = "unescaped-quotes.csv"
+  private val carsFile = "test-data/cars.csv"
+  private val carsMalformedFile = "test-data/cars-malformed.csv"
+  private val carsFile8859 = "test-data/cars_iso-8859-1.csv"
+  private val carsTsvFile = "test-data/cars.tsv"
+  private val carsAltFile = "test-data/cars-alternative.csv"
+  private val carsUnbalancedQuotesFile = "test-data/cars-unbalanced-quotes.csv"
+  private val carsNullFile = "test-data/cars-null.csv"
+  private val carsBlankColName = "test-data/cars-blank-column-name.csv"
+  private val emptyFile = "test-data/empty.csv"
+  private val commentsFile = "test-data/comments.csv"
+  private val disableCommentsFile = "test-data/disable_comments.csv"
+  private val boolFile = "test-data/bool.csv"
+  private val decimalFile = "test-data/decimal.csv"
+  private val simpleSparseFile = "test-data/simple_sparse.csv"
+  private val numbersFile = "test-data/numbers.csv"
+  private val datesFile = "test-data/dates.csv"
+  private val unescapedQuotesFile = "test-data/unescaped-quotes.csv"
 
   private def testFile(fileName: String): String = {
     Thread.currentThread().getContextClassLoader.getResource(fileName).toString
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
index 0f74094699abd465976d2ada28b68ac2aa31eade..4aa046bd91e0284f19ada9d63cfb25366d6cf633 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetIOSuite.scala
@@ -568,7 +568,7 @@ class ParquetIOSuite extends QueryTest with ParquetTest with SharedSQLContext {
       withSQLConf(SQLConf.PARQUET_VECTORIZED_READER_ENABLED.key -> vectorized) {
         checkAnswer(
           // Decimal column in this file is encoded using plain dictionary
-          readResourceParquetFile("dec-in-i32.parquet"),
+          readResourceParquetFile("test-data/dec-in-i32.parquet"),
           spark.range(1 << 4).select('id % 10 cast DecimalType(5, 2) as 'i32_dec))
       }
     }
@@ -579,7 +579,7 @@ class ParquetIOSuite extends QueryTest with ParquetTest with SharedSQLContext {
       withSQLConf(SQLConf.PARQUET_VECTORIZED_READER_ENABLED.key -> vectorized) {
         checkAnswer(
           // Decimal column in this file is encoded using plain dictionary
-          readResourceParquetFile("dec-in-i64.parquet"),
+          readResourceParquetFile("test-data/dec-in-i64.parquet"),
           spark.range(1 << 4).select('id % 10 cast DecimalType(10, 2) as 'i64_dec))
       }
     }
@@ -590,7 +590,7 @@ class ParquetIOSuite extends QueryTest with ParquetTest with SharedSQLContext {
       withSQLConf(SQLConf.PARQUET_VECTORIZED_READER_ENABLED.key -> vectorized) {
         checkAnswer(
           // Decimal column in this file is encoded using plain dictionary
-          readResourceParquetFile("dec-in-fixed-len.parquet"),
+          readResourceParquetFile("test-data/dec-in-fixed-len.parquet"),
           spark.range(1 << 4).select('id % 10 cast DecimalType(10, 2) as 'fixed_len_dec))
       }
     }
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetProtobufCompatibilitySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetProtobufCompatibilitySuite.scala
index 98333e58cada80800851c7f02ee76dbca266ccf7..fa88019298a6962429f91ab9505d700699619530 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetProtobufCompatibilitySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetProtobufCompatibilitySuite.scala
@@ -22,12 +22,12 @@ import org.apache.spark.sql.test.SharedSQLContext
 
 class ParquetProtobufCompatibilitySuite extends ParquetCompatibilityTest with SharedSQLContext {
   test("unannotated array of primitive type") {
-    checkAnswer(readResourceParquetFile("old-repeated-int.parquet"), Row(Seq(1, 2, 3)))
+    checkAnswer(readResourceParquetFile("test-data/old-repeated-int.parquet"), Row(Seq(1, 2, 3)))
   }
 
   test("unannotated array of struct") {
     checkAnswer(
-      readResourceParquetFile("old-repeated-message.parquet"),
+      readResourceParquetFile("test-data/old-repeated-message.parquet"),
       Row(
         Seq(
           Row("First inner", null, null),
@@ -35,14 +35,14 @@ class ParquetProtobufCompatibilitySuite extends ParquetCompatibilityTest with Sh
           Row(null, null, "Third inner"))))
 
     checkAnswer(
-      readResourceParquetFile("proto-repeated-struct.parquet"),
+      readResourceParquetFile("test-data/proto-repeated-struct.parquet"),
       Row(
         Seq(
           Row("0 - 1", "0 - 2", "0 - 3"),
           Row("1 - 1", "1 - 2", "1 - 3"))))
 
     checkAnswer(
-      readResourceParquetFile("proto-struct-with-array-many.parquet"),
+      readResourceParquetFile("test-data/proto-struct-with-array-many.parquet"),
       Seq(
         Row(
           Seq(
@@ -60,13 +60,13 @@ class ParquetProtobufCompatibilitySuite extends ParquetCompatibilityTest with Sh
 
   test("struct with unannotated array") {
     checkAnswer(
-      readResourceParquetFile("proto-struct-with-array.parquet"),
+      readResourceParquetFile("test-data/proto-struct-with-array.parquet"),
       Row(10, 9, Seq.empty, null, Row(9), Seq(Row(9), Row(10))))
   }
 
   test("unannotated array of struct with unannotated array") {
     checkAnswer(
-      readResourceParquetFile("nested-array-struct.parquet"),
+      readResourceParquetFile("test-data/nested-array-struct.parquet"),
       Seq(
         Row(2, Seq(Row(1, Seq(Row(3))))),
         Row(5, Seq(Row(4, Seq(Row(6))))),
@@ -75,7 +75,7 @@ class ParquetProtobufCompatibilitySuite extends ParquetCompatibilityTest with Sh
 
   test("unannotated array of string") {
     checkAnswer(
-      readResourceParquetFile("proto-repeated-string.parquet"),
+      readResourceParquetFile("test-data/proto-repeated-string.parquet"),
       Seq(
         Row(Seq("hello", "world")),
         Row(Seq("good", "bye")),
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetThriftCompatibilitySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetThriftCompatibilitySuite.scala
index ff5706999a6dd85920affa675d76b08a91bc6e52..4157a5b46dc4225b058b8aa1d50d096efe8da6ce 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetThriftCompatibilitySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetThriftCompatibilitySuite.scala
@@ -23,8 +23,8 @@ import org.apache.spark.sql.test.SharedSQLContext
 class ParquetThriftCompatibilitySuite extends ParquetCompatibilityTest with SharedSQLContext {
   import ParquetCompatibilityTest._
 
-  private val parquetFilePath =
-    Thread.currentThread().getContextClassLoader.getResource("parquet-thrift-compat.snappy.parquet")
+  private val parquetFilePath = Thread.currentThread().getContextClassLoader.getResource(
+    "test-data/parquet-thrift-compat.snappy.parquet")
 
   test("Read Parquet file generated by parquet-thrift") {
     logInfo(
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala
index 71d3da915840a21094544ff36d5af3f4d2c66e35..d11c2acb815d4618ec718ea045f5f1fc324641d8 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/text/TextSuite.scala
@@ -66,7 +66,7 @@ class TextSuite extends QueryTest with SharedSQLContext {
 
   test("reading partitioned data using read.textFile()") {
     val partitionedData = Thread.currentThread().getContextClassLoader
-      .getResource("text-partitioned").toString
+      .getResource("test-data/text-partitioned").toString
     val ds = spark.read.textFile(partitionedData)
     val data = ds.collect()
 
@@ -76,7 +76,7 @@ class TextSuite extends QueryTest with SharedSQLContext {
 
   test("support for partitioned reading using read.text()") {
     val partitionedData = Thread.currentThread().getContextClassLoader
-      .getResource("text-partitioned").toString
+      .getResource("test-data/text-partitioned").toString
     val df = spark.read.text(partitionedData)
     val data = df.filter("year = '2015'").select("value").collect()
 
@@ -155,7 +155,7 @@ class TextSuite extends QueryTest with SharedSQLContext {
   }
 
   private def testFile: String = {
-    Thread.currentThread().getContextClassLoader.getResource("text-suite.txt").toString
+    Thread.currentThread().getContextClassLoader.getResource("test-data/text-suite.txt").toString
   }
 
   /** Verifies data and schema. */