I would like to register a UDAF class (written in Scala or Python) and use it in Spark SQL.
For example:
Mock code
// mock code:
class MyUDAF extends UserDefinedAggregateFunction{
...
}
spark.udaf.registerJavaFunction("myagg", "MyUDAF",IntegerType)
Then I can use the udaf directly within spark sql like following:
spark.sql("select myagg(field) from mytable group by something")
Spark has only provided spark.udf.registerJavaFunction method to register a UDF class.
Anyone knows how to register a UDAF?