pyspark.sql.functions.grouping#
- pyspark.sql.functions.grouping(col)[source]#
Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.
New in version 2.0.0.
Changed in version 3.4.0: Supports Spark Connect.
- Parameters
- col
Column
or str column to check if it’s aggregated.
- col
- Returns
Column
returns 1 for aggregated or 0 for not aggregated in the result set.
Examples
>>> df = spark.createDataFrame([("Alice", 2), ("Bob", 5)], ("name", "age")) >>> df.cube("name").agg(grouping("name"), sum("age")).orderBy("name").show() +-----+--------------+--------+ | name|grouping(name)|sum(age)| +-----+--------------+--------+ | NULL| 1| 7| |Alice| 0| 2| | Bob| 0| 5| +-----+--------------+--------+