Example:
squared_udf = udf(squared, IntegerType()) df.withColumn("squared_val", squared_udf(df.value))
# Read df = spark.read.option("header", "true").csv("path/to/file.csv") df.write.parquet("output.parquet") 4.2 Common Transformations | Operation | Example | |------------------|-------------------------------------------| | Select columns | df.select("name", "age") | | Filter rows | df.filter(df.age > 21) | | Add column | df.withColumn("new", df.value * 2) | | Group and aggregate | df.groupBy("dept").avg("salary") | | Join | df1.join(df2, "id", "inner") | 4.3 Handling Missing Data df.dropna(how="any", subset=["important_col"]) df.fillna("age": 0, "name": "unknown") 4.4 User‑Defined Functions (UDFs) When built‑in functions are insufficient:
spark.stop()
from pyspark.sql.functions import udf def squared(x): return x * x
Example:
squared_udf = udf(squared, IntegerType()) df.withColumn("squared_val", squared_udf(df.value)) beginning apache spark 3 pdf
# Read df = spark.read.option("header", "true").csv("path/to/file.csv") df.write.parquet("output.parquet") 4.2 Common Transformations | Operation | Example | |------------------|-------------------------------------------| | Select columns | df.select("name", "age") | | Filter rows | df.filter(df.age > 21) | | Add column | df.withColumn("new", df.value * 2) | | Group and aggregate | df.groupBy("dept").avg("salary") | | Join | df1.join(df2, "id", "inner") | 4.3 Handling Missing Data df.dropna(how="any", subset=["important_col"]) df.fillna("age": 0, "name": "unknown") 4.4 User‑Defined Functions (UDFs) When built‑in functions are insufficient: Example: squared_udf = udf(squared, IntegerType()) df
spark.stop()
from pyspark.sql.functions import udf def squared(x): return x * x Example: squared_udf = udf(squared