Rate This Document
Findability
Accuracy
Completeness
Readability

Deploying the Spark UDF Plugin

  1. Add the following environment variable to the ~/.bashrc file on all nodes:
    export OMNI_HOME=/opt/omni-operator
  2. Place the JAR package on which UDFs depend to the /user/hive-udf directory of HDFS.
    • The /user/hive-udf directory can be customized.
    • The JAR package on which UDF depends needs to be provided by yourself.
  3. Register Hive UDF on the management nodes of the cluster.

    For details about how to register it, see Integration with Hive UDFs/UDAFs/UDTFs.