Is there any difference in semantics between df.na().drop()
and df.filter(df.col("onlyColumnInOneColumnDataFrame").isNotNull() && !df.col("onlyColumnInOneColumnDataFrame").isNaN())
where df
is Apache Spark Dataframe
?
Or shall I consider it as a bug if the first one does NOT return afterwards null
(not a String null, but simply a null
value) in the column onlyColumnInOneColumnDataFrame
and the second one does?
EDIT: added !isNaN()
as well. The onlyColumnInOneColumnDataFrame
is the only column in the given Dataframe
. Let's say it's type is Integer
.