Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions .github/workflows/pr_build_linux.yml
Original file line number Diff line number Diff line change
Expand Up @@ -356,6 +356,7 @@ jobs:
org.apache.comet.exec.CometJoinSuite
org.apache.comet.CometNativeSuite
org.apache.comet.CometSparkSessionExtensionsSuite
org.apache.comet.CometUserUdfSuite
org.apache.spark.CometPluginsSuite
org.apache.spark.CometPluginsDefaultSuite
org.apache.spark.CometPluginsNonOverrideSuite
Expand Down
1 change: 1 addition & 0 deletions .github/workflows/pr_build_macos.yml
Original file line number Diff line number Diff line change
Expand Up @@ -195,6 +195,7 @@ jobs:
org.apache.comet.exec.CometJoinSuite
org.apache.comet.CometNativeSuite
org.apache.comet.CometSparkSessionExtensionsSuite
org.apache.comet.CometUserUdfSuite
org.apache.spark.CometPluginsSuite
org.apache.spark.CometPluginsDefaultSuite
org.apache.spark.CometPluginsNonOverrideSuite
Expand Down
141 changes: 141 additions & 0 deletions common/src/main/java/org/apache/comet/udf/CometUdfBridge.java
Original file line number Diff line number Diff line change
@@ -0,0 +1,141 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/

package org.apache.comet.udf;

import java.util.LinkedHashMap;
import java.util.Map;

import org.apache.arrow.c.ArrowArray;
import org.apache.arrow.c.ArrowSchema;
import org.apache.arrow.c.Data;
import org.apache.arrow.memory.BufferAllocator;
import org.apache.arrow.vector.FieldVector;
import org.apache.arrow.vector.ValueVector;

/**
* JNI entry point for native execution to invoke a {@link CometUDF}. Matches the static-method
* pattern used by CometScalarSubquery so the native side can dispatch via
* call_static_method_unchecked.
*/
public class CometUdfBridge {

// Per-thread, bounded LRU of UDF instances keyed by class name. Comet
// native execution threads (Tokio/DataFusion worker pool) are reused
// across tasks within an executor, so the effective lifetime of cached
// entries is the worker thread (i.e. the executor JVM). This is fine for
// stateless UDFs like ArrayExistsUDF; future stateful UDFs would need
// explicit per-task isolation.
private static final int CACHE_CAPACITY = 64;

private static final ThreadLocal<LinkedHashMap<String, CometUDF>> INSTANCES =
ThreadLocal.withInitial(
() ->
new LinkedHashMap<String, CometUDF>(CACHE_CAPACITY, 0.75f, true) {
@Override
protected boolean removeEldestEntry(Map.Entry<String, CometUDF> eldest) {
return size() > CACHE_CAPACITY;
}
});

/**
* Called from native via JNI.
*
* @param udfClassName fully-qualified class name implementing CometUDF
* @param inputArrayPtrs addresses of pre-allocated FFI_ArrowArray structs (one per input)
* @param inputSchemaPtrs addresses of pre-allocated FFI_ArrowSchema structs (one per input)
* @param outArrayPtr address of pre-allocated FFI_ArrowArray for the result
* @param outSchemaPtr address of pre-allocated FFI_ArrowSchema for the result
*/
public static void evaluate(
String udfClassName,
long[] inputArrayPtrs,
long[] inputSchemaPtrs,
long outArrayPtr,
long outSchemaPtr) {
LinkedHashMap<String, CometUDF> cache = INSTANCES.get();
CometUDF udf = cache.get(udfClassName);
if (udf == null) {
try {
// Resolve via the executor's context classloader so user-supplied UDF jars
// (added via spark.jars / --jars) are visible.
ClassLoader cl = Thread.currentThread().getContextClassLoader();
if (cl == null) {
cl = CometUdfBridge.class.getClassLoader();
}
udf =
(CometUDF) Class.forName(udfClassName, true, cl).getDeclaredConstructor().newInstance();
} catch (ReflectiveOperationException e) {
throw new RuntimeException("Failed to instantiate CometUDF: " + udfClassName, e);
}
cache.put(udfClassName, udf);
}

BufferAllocator allocator = org.apache.comet.package$.MODULE$.CometArrowAllocator();

ValueVector[] inputs = new ValueVector[inputArrayPtrs.length];
ValueVector result = null;
try {
for (int i = 0; i < inputArrayPtrs.length; i++) {
ArrowArray inArr = ArrowArray.wrap(inputArrayPtrs[i]);
ArrowSchema inSch = ArrowSchema.wrap(inputSchemaPtrs[i]);
inputs[i] = Data.importVector(allocator, inArr, inSch, null);
}

result = udf.evaluate(inputs);
if (!(result instanceof FieldVector)) {
throw new RuntimeException(
"CometUDF.evaluate() must return a FieldVector, got: " + result.getClass().getName());
}
// Result length must match the longest input. Scalar (length-1) inputs
// are allowed to be shorter, but a vector input bounds the output.
int expectedLen = 0;
for (ValueVector v : inputs) {
expectedLen = Math.max(expectedLen, v.getValueCount());
}
if (result.getValueCount() != expectedLen) {
throw new RuntimeException(
"CometUDF.evaluate() returned "
+ result.getValueCount()
+ " rows, expected "
+ expectedLen);
}
ArrowArray outArr = ArrowArray.wrap(outArrayPtr);
ArrowSchema outSch = ArrowSchema.wrap(outSchemaPtr);
Data.exportVector(allocator, (FieldVector) result, null, outArr, outSch);
} finally {
for (ValueVector v : inputs) {
if (v != null) {
try {
v.close();
} catch (RuntimeException ignored) {
// do not mask the original throwable
}
}
}
if (result != null) {
try {
result.close();
} catch (RuntimeException ignored) {
// do not mask the original throwable
}
}
}
}
}
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/

package org.apache.comet.udf

import java.util.UUID
import java.util.concurrent.ConcurrentHashMap

import org.apache.spark.sql.catalyst.expressions.Expression

/**
* Thread-safe registry bridging plan-time Spark expressions to execution-time UDF lookup. At plan
* time the serde layer registers a lambda expression under a unique key; at execution time the
* UDF retrieves it by that key (passed as a scalar argument).
*/
object CometLambdaRegistry {

private val registry = new ConcurrentHashMap[String, Expression]()

def register(expression: Expression): String = {
val key = UUID.randomUUID().toString
registry.put(key, expression)
key
}

def get(key: String): Expression = {
val expr = registry.get(key)
if (expr == null) {
throw new IllegalStateException(
s"Lambda expression not found in registry for key: $key. " +
"This indicates a lifecycle issue between plan creation and execution.")
}
expr
}

def remove(key: String): Unit = {
registry.remove(key)
}

// Visible for testing
def size(): Int = registry.size()
}
55 changes: 55 additions & 0 deletions common/src/main/scala/org/apache/comet/udf/CometUDF.scala
Original file line number Diff line number Diff line change
@@ -0,0 +1,55 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/

package org.apache.comet.udf

import org.apache.arrow.vector.ValueVector
import org.apache.spark.sql.types.DataType

/**
* Scalar UDF invoked from native execution via JNI. Receives Arrow vectors as input and returns
* an Arrow vector.
*
* - Vector arguments arrive at the row count of the current batch.
* - Scalar (literal-folded) arguments arrive as length-1 vectors and must be read at index 0.
* - The returned vector's length must match the longest input.
*
* Implementations must have a public no-arg constructor and should be stateless: instances are
* cached per executor thread for the lifetime of the JVM.
*/
trait CometUDF {

/** UDF name as invoked from SQL or DataFrame. Must match the name registered with Spark. */
def name: String

/** Output Arrow vector type. */
def returnType: DataType

/** Whether the result vector may contain nulls. */
def nullable: Boolean = true

/**
* Input data types. Required only for columnar-only registration via
* [[CometUdfRegistry.registerColumnarOnly]]; ignored when a row-based Spark UDF is also
* registered (Spark uses its own input schema in that case).
*/
def inputTypes: Seq[DataType] = Seq.empty

def evaluate(inputs: Array[ValueVector]): ValueVector
}
Loading
Loading