pyspark.sql.DataFrame.exceptAll¶
-
DataFrame.
exceptAll
(other)[source]¶ Return a new
DataFrame
containing rows in thisDataFrame
but not in anotherDataFrame
while preserving duplicates.This is equivalent to EXCEPT ALL in SQL.
>>> df1 = spark.createDataFrame( ... [("a", 1), ("a", 1), ("a", 1), ("a", 2), ("b", 3), ("c", 4)], ["C1", "C2"]) >>> df2 = spark.createDataFrame([("a", 1), ("b", 3)], ["C1", "C2"])
>>> df1.exceptAll(df2).show() +---+---+ | C1| C2| +---+---+ | a| 1| | a| 1| | a| 2| | c| 4| +---+---+
Also as standard in SQL, this function resolves columns by position (not by name).
New in version 2.4.