pyspark.sql.DataFrame.exceptAll#
- DataFrame.exceptAll(other)[source]#
Return a new
DataFrame
containing rows in thisDataFrame
but not in anotherDataFrame
while preserving duplicates.This is equivalent to EXCEPT ALL in SQL. As standard in SQL, this function resolves columns by position (not by name).
New in version 2.4.0.
Changed in version 3.4.0: Supports Spark Connect.
Examples
>>> df1 = spark.createDataFrame( ... [("a", 1), ("a", 1), ("a", 1), ("a", 2), ("b", 3), ("c", 4)], ["C1", "C2"]) >>> df2 = spark.createDataFrame([("a", 1), ("b", 3)], ["C1", "C2"]) >>> df1.exceptAll(df2).show() +---+---+ | C1| C2| +---+---+ | a| 1| | a| 1| | a| 2| | c| 4| +---+---+