diff --git a/python/pyspark/sql/readwriter.py b/python/pyspark/sql/readwriter.py index c4134b7bf690b..31c1f2f7ca3c5 100644 --- a/python/pyspark/sql/readwriter.py +++ b/python/pyspark/sql/readwriter.py @@ -728,7 +728,7 @@ def csv(self, path, schema=None, sep=None, encoding=None, quote=None, escape=Non * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the input, consider the value as an unquoted value. This will make the parser accumulate all characters until the delimiter or a line ending is found in the input. - * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the input, the content parsed + * ``SKIP_VALUE``: If unescaped quotes are found in the input, the content parsed for the given value will be skipped and the value set in nullValue will be produced instead. * ``RAISE_ERROR``: If unescaped quotes are found in the input, a TextParsingException diff --git a/python/pyspark/sql/streaming.py b/python/pyspark/sql/streaming.py index 8b8484defb600..2c90d7f2dee70 100644 --- a/python/pyspark/sql/streaming.py +++ b/python/pyspark/sql/streaming.py @@ -941,7 +941,7 @@ def csv(self, path, schema=None, sep=None, encoding=None, quote=None, escape=Non * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the input, consider the value as an unquoted value. This will make the parser accumulate all characters until the delimiter or a line ending is found in the input. - * ``STOP_AT_DELIMITER``: If unescaped quotes are found in the input, the content parsed + * ``SKIP_VALUE``: If unescaped quotes are found in the input, the content parsed for the given value will be skipped and the value set in nullValue will be produced instead. * ``RAISE_ERROR``: If unescaped quotes are found in the input, a TextParsingException diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala index 9ff37a6f1c69c..f7e1903da6877 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala @@ -747,7 +747,7 @@ class DataFrameReader private[sql](sparkSession: SparkSession) extends Logging { *