Skip to content
Snippets Groups Projects
Commit a51d51ff authored by Patrick Wendell's avatar Patrick Wendell
Browse files

SPARK-5850: Remove experimental label for Scala 2.11 and FlumePollingStream

Author: Patrick Wendell <patrick@databricks.com>

Closes #4638 from pwendell/SPARK-5850 and squashes the following commits:

386126f [Patrick Wendell] SPARK-5850: Remove experimental label for Scala 2.11 and FlumePollingStream.
parent ac6fe67e
No related branches found
No related tags found
No related merge requests found
...@@ -111,9 +111,9 @@ To produce a Spark package compiled with Scala 2.11, use the `-Dscala-2.11` prop ...@@ -111,9 +111,9 @@ To produce a Spark package compiled with Scala 2.11, use the `-Dscala-2.11` prop
dev/change-version-to-2.11.sh dev/change-version-to-2.11.sh
mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package
Scala 2.11 support in Spark is experimental and does not support a few features. Scala 2.11 support in Spark does not support a few features due to dependencies
Specifically, Spark's external Kafka library and JDBC component are not yet which are themselves not Scala 2.11 ready. Specifically, Spark's external
supported in Scala 2.11 builds. Kafka library and JDBC component are not yet supported in Scala 2.11 builds.
# Spark Tests in Maven # Spark Tests in Maven
......
...@@ -64,7 +64,7 @@ configuring Flume agents. ...@@ -64,7 +64,7 @@ configuring Flume agents.
3. **Deploying:** Package `spark-streaming-flume_{{site.SCALA_BINARY_VERSION}}` and its dependencies (except `spark-core_{{site.SCALA_BINARY_VERSION}}` and `spark-streaming_{{site.SCALA_BINARY_VERSION}}` which are provided by `spark-submit`) into the application JAR. Then use `spark-submit` to launch your application (see [Deploying section](streaming-programming-guide.html#deploying-applications) in the main programming guide). 3. **Deploying:** Package `spark-streaming-flume_{{site.SCALA_BINARY_VERSION}}` and its dependencies (except `spark-core_{{site.SCALA_BINARY_VERSION}}` and `spark-streaming_{{site.SCALA_BINARY_VERSION}}` which are provided by `spark-submit`) into the application JAR. Then use `spark-submit` to launch your application (see [Deploying section](streaming-programming-guide.html#deploying-applications) in the main programming guide).
## Approach 2 (Experimental): Pull-based Approach using a Custom Sink ## Approach 2: Pull-based Approach using a Custom Sink
Instead of Flume pushing data directly to Spark Streaming, this approach runs a custom Flume sink that allows the following. Instead of Flume pushing data directly to Spark Streaming, this approach runs a custom Flume sink that allows the following.
- Flume pushes data into the sink, and the data stays buffered. - Flume pushes data into the sink, and the data stays buffered.
......
...@@ -19,7 +19,6 @@ package org.apache.spark.streaming.flume ...@@ -19,7 +19,6 @@ package org.apache.spark.streaming.flume
import java.net.InetSocketAddress import java.net.InetSocketAddress
import org.apache.spark.annotation.Experimental
import org.apache.spark.storage.StorageLevel import org.apache.spark.storage.StorageLevel
import org.apache.spark.streaming.StreamingContext import org.apache.spark.streaming.StreamingContext
import org.apache.spark.streaming.api.java.{JavaReceiverInputDStream, JavaStreamingContext} import org.apache.spark.streaming.api.java.{JavaReceiverInputDStream, JavaStreamingContext}
...@@ -121,7 +120,6 @@ object FlumeUtils { ...@@ -121,7 +120,6 @@ object FlumeUtils {
* @param port Port of the host at which the Spark Sink is listening * @param port Port of the host at which the Spark Sink is listening
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
ssc: StreamingContext, ssc: StreamingContext,
hostname: String, hostname: String,
...@@ -138,7 +136,6 @@ object FlumeUtils { ...@@ -138,7 +136,6 @@ object FlumeUtils {
* @param addresses List of InetSocketAddresses representing the hosts to connect to. * @param addresses List of InetSocketAddresses representing the hosts to connect to.
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
ssc: StreamingContext, ssc: StreamingContext,
addresses: Seq[InetSocketAddress], addresses: Seq[InetSocketAddress],
...@@ -159,7 +156,6 @@ object FlumeUtils { ...@@ -159,7 +156,6 @@ object FlumeUtils {
* result in this stream using more threads * result in this stream using more threads
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
ssc: StreamingContext, ssc: StreamingContext,
addresses: Seq[InetSocketAddress], addresses: Seq[InetSocketAddress],
...@@ -178,7 +174,6 @@ object FlumeUtils { ...@@ -178,7 +174,6 @@ object FlumeUtils {
* @param hostname Hostname of the host on which the Spark Sink is running * @param hostname Hostname of the host on which the Spark Sink is running
* @param port Port of the host at which the Spark Sink is listening * @param port Port of the host at which the Spark Sink is listening
*/ */
@Experimental
def createPollingStream( def createPollingStream(
jssc: JavaStreamingContext, jssc: JavaStreamingContext,
hostname: String, hostname: String,
...@@ -195,7 +190,6 @@ object FlumeUtils { ...@@ -195,7 +190,6 @@ object FlumeUtils {
* @param port Port of the host at which the Spark Sink is listening * @param port Port of the host at which the Spark Sink is listening
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
jssc: JavaStreamingContext, jssc: JavaStreamingContext,
hostname: String, hostname: String,
...@@ -212,7 +206,6 @@ object FlumeUtils { ...@@ -212,7 +206,6 @@ object FlumeUtils {
* @param addresses List of InetSocketAddresses on which the Spark Sink is running. * @param addresses List of InetSocketAddresses on which the Spark Sink is running.
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
jssc: JavaStreamingContext, jssc: JavaStreamingContext,
addresses: Array[InetSocketAddress], addresses: Array[InetSocketAddress],
...@@ -233,7 +226,6 @@ object FlumeUtils { ...@@ -233,7 +226,6 @@ object FlumeUtils {
* result in this stream using more threads * result in this stream using more threads
* @param storageLevel Storage level to use for storing the received objects * @param storageLevel Storage level to use for storing the received objects
*/ */
@Experimental
def createPollingStream( def createPollingStream(
jssc: JavaStreamingContext, jssc: JavaStreamingContext,
addresses: Array[InetSocketAddress], addresses: Array[InetSocketAddress],
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment