From 68ef61bb656bd9c08239726913ca8ab271d52786 Mon Sep 17 00:00:00 2001
From: Chris Snow <chsnow123@gmail.com>
Date: Thu, 12 Nov 2015 15:50:47 -0800
Subject: [PATCH] [SPARK-11658] simplify documentation for PySpark combineByKey

Author: Chris Snow <chsnow123@gmail.com>

Closes #9640 from snowch/patch-3.
---
 python/pyspark/rdd.py | 1 -
 1 file changed, 1 deletion(-)

diff --git a/python/pyspark/rdd.py b/python/pyspark/rdd.py
index 56e892243c..4b4d59647b 100644
--- a/python/pyspark/rdd.py
+++ b/python/pyspark/rdd.py
@@ -1760,7 +1760,6 @@ class RDD(object):
         In addition, users can control the partitioning of the output RDD.
 
         >>> x = sc.parallelize([("a", 1), ("b", 1), ("a", 1)])
-        >>> def f(x): return x
         >>> def add(a, b): return a + str(b)
         >>> sorted(x.combineByKey(str, add, add).collect())
         [('a', '11'), ('b', '1')]
-- 
GitLab