public class tritonserver extends tritonserver
Modifier and Type | Field and Description |
---|---|
static int |
TRITONBACKEND_API_VERSION_MAJOR
TRITONBACKEND API Version
The TRITONBACKEND API is versioned with major and minor version
numbers.
|
static int |
TRITONBACKEND_API_VERSION_MINOR |
static int |
TRITONBACKEND_ARTIFACT_FILESYSTEM
enum TRITONBACKEND_ArtifactType
|
static int |
TRITONBACKEND_EXECUTION_BLOCKING
enum TRITONBACKEND_ExecutionPolicy
|
static int |
TRITONBACKEND_EXECUTION_DEVICE_BLOCKING
enum TRITONBACKEND_ExecutionPolicy
|
static int |
TRITONREPOAGENT_ACTION_LOAD
enum TRITONREPOAGENT_ActionType
|
static int |
TRITONREPOAGENT_ACTION_LOAD_COMPLETE
enum TRITONREPOAGENT_ActionType
|
static int |
TRITONREPOAGENT_ACTION_LOAD_FAIL
enum TRITONREPOAGENT_ActionType
|
static int |
TRITONREPOAGENT_ACTION_UNLOAD
enum TRITONREPOAGENT_ActionType
|
static int |
TRITONREPOAGENT_ACTION_UNLOAD_COMPLETE
enum TRITONREPOAGENT_ActionType
|
static int |
TRITONREPOAGENT_API_VERSION_MAJOR
TRITONREPOAGENT API Version
The TRITONREPOAGENT API is versioned with major and minor version
numbers.
|
static int |
TRITONREPOAGENT_API_VERSION_MINOR |
static int |
TRITONREPOAGENT_ARTIFACT_FILESYSTEM
enum TRITONREPOAGENT_ArtifactType
|
static int |
TRITONREPOAGENT_ARTIFACT_REMOTE_FILESYSTEM
enum TRITONREPOAGENT_ArtifactType
|
static int |
TRITONSERVER_API_VERSION_MAJOR
TRITONSERVER API Version
The TRITONSERVER API is versioned with major and minor version
numbers.
|
static int |
TRITONSERVER_API_VERSION_MINOR |
static int |
TRITONSERVER_BATCH_FIRST_DIM
enum TRITONSERVER_ModelBatchFlag
|
static int |
TRITONSERVER_BATCH_UNKNOWN
enum TRITONSERVER_ModelBatchFlag
|
static int |
TRITONSERVER_ERROR_ALREADY_EXISTS
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_CANCELLED
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_INTERNAL
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_INVALID_ARG
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_NOT_FOUND
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_UNAVAILABLE
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_UNKNOWN
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_ERROR_UNSUPPORTED
enum TRITONSERVER_Error_Code
|
static int |
TRITONSERVER_INDEX_FLAG_READY
enum TRITONSERVER_ModelIndexFlag
|
static int |
TRITONSERVER_INSTANCEGROUPKIND_AUTO
enum TRITONSERVER_InstanceGroupKind
|
static int |
TRITONSERVER_INSTANCEGROUPKIND_CPU
enum TRITONSERVER_InstanceGroupKind
|
static int |
TRITONSERVER_INSTANCEGROUPKIND_GPU
enum TRITONSERVER_InstanceGroupKind
|
static int |
TRITONSERVER_INSTANCEGROUPKIND_MODEL
enum TRITONSERVER_InstanceGroupKind
|
static int |
TRITONSERVER_LOG_DEFAULT
enum TRITONSERVER_LogFormat
|
static int |
TRITONSERVER_LOG_ERROR
enum TRITONSERVER_LogLevel
|
static int |
TRITONSERVER_LOG_INFO
enum TRITONSERVER_LogLevel
|
static int |
TRITONSERVER_LOG_ISO8601
enum TRITONSERVER_LogFormat
|
static int |
TRITONSERVER_LOG_VERBOSE
enum TRITONSERVER_LogLevel
|
static int |
TRITONSERVER_LOG_WARN
enum TRITONSERVER_LogLevel
|
static int |
TRITONSERVER_MEMORY_CPU
enum TRITONSERVER_MemoryType
|
static int |
TRITONSERVER_MEMORY_CPU_PINNED
enum TRITONSERVER_MemoryType
|
static int |
TRITONSERVER_MEMORY_GPU
enum TRITONSERVER_MemoryType
|
static int |
TRITONSERVER_METRIC_KIND_COUNTER
enum TRITONSERVER_MetricKind
|
static int |
TRITONSERVER_METRIC_KIND_GAUGE
enum TRITONSERVER_MetricKind
|
static int |
TRITONSERVER_METRIC_PROMETHEUS
enum TRITONSERVER_MetricFormat
|
static int |
TRITONSERVER_MODEL_CONTROL_EXPLICIT
enum TRITONSERVER_ModelControlMode
|
static int |
TRITONSERVER_MODEL_CONTROL_NONE
enum TRITONSERVER_ModelControlMode
|
static int |
TRITONSERVER_MODEL_CONTROL_POLL
enum TRITONSERVER_ModelControlMode
|
static int |
TRITONSERVER_PARAMETER_BOOL
enum TRITONSERVER_ParameterType
|
static int |
TRITONSERVER_PARAMETER_BYTES
enum TRITONSERVER_ParameterType
|
static int |
TRITONSERVER_PARAMETER_INT
enum TRITONSERVER_ParameterType
|
static int |
TRITONSERVER_PARAMETER_STRING
enum TRITONSERVER_ParameterType
|
static int |
TRITONSERVER_RATE_LIMIT_EXEC_COUNT
enum TRITONSERVER_RateLimitMode
|
static int |
TRITONSERVER_RATE_LIMIT_OFF
enum TRITONSERVER_RateLimitMode
|
static int |
TRITONSERVER_REQUEST_FLAG_SEQUENCE_END
enum TRITONSERVER_RequestFlag
|
static int |
TRITONSERVER_REQUEST_FLAG_SEQUENCE_START
enum TRITONSERVER_RequestFlag
|
static int |
TRITONSERVER_REQUEST_RELEASE_ALL
enum TRITONSERVER_RequestReleaseFlag
|
static int |
TRITONSERVER_REQUEST_RELEASE_RESCHEDULE
enum TRITONSERVER_RequestReleaseFlag
|
static int |
TRITONSERVER_RESPONSE_COMPLETE_FINAL
enum TRITONSERVER_ResponseCompleteFlag
|
static int |
TRITONSERVER_TRACE_COMPUTE_END
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_COMPUTE_INPUT_END
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_COMPUTE_OUTPUT_START
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_COMPUTE_START
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_LEVEL_DISABLED
enum TRITONSERVER_InferenceTraceLevel
|
static int |
TRITONSERVER_TRACE_LEVEL_MAX
enum TRITONSERVER_InferenceTraceLevel
|
static int |
TRITONSERVER_TRACE_LEVEL_MIN
enum TRITONSERVER_InferenceTraceLevel
|
static int |
TRITONSERVER_TRACE_LEVEL_TENSORS
enum TRITONSERVER_InferenceTraceLevel
|
static int |
TRITONSERVER_TRACE_LEVEL_TIMESTAMPS
enum TRITONSERVER_InferenceTraceLevel
|
static int |
TRITONSERVER_TRACE_QUEUE_START
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_REQUEST_END
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_REQUEST_START
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_TENSOR_BACKEND_INPUT
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_TENSOR_BACKEND_OUTPUT
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TRACE_TENSOR_QUEUE_INPUT
enum TRITONSERVER_InferenceTraceActivity
|
static int |
TRITONSERVER_TXN_DECOUPLED
enum TRITONSERVER_ModelTxnPropertyFlag
|
static int |
TRITONSERVER_TXN_ONE_TO_ONE
enum TRITONSERVER_ModelTxnPropertyFlag
|
static int |
TRITONSERVER_TYPE_BF16
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_BOOL
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_BYTES
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_FP16
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_FP32
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_FP64
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_INT16
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_INT32
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_INT64
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_INT8
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_INVALID
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_UINT16
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_UINT32
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_UINT64
enum TRITONSERVER_DataType
|
static int |
TRITONSERVER_TYPE_UINT8
enum TRITONSERVER_DataType
|
Constructor and Description |
---|
tritonserver() |
map
public static final int TRITONSERVER_API_VERSION_MAJOR
public static final int TRITONSERVER_API_VERSION_MINOR
public static final int TRITONSERVER_TYPE_INVALID
public static final int TRITONSERVER_TYPE_BOOL
public static final int TRITONSERVER_TYPE_UINT8
public static final int TRITONSERVER_TYPE_UINT16
public static final int TRITONSERVER_TYPE_UINT32
public static final int TRITONSERVER_TYPE_UINT64
public static final int TRITONSERVER_TYPE_INT8
public static final int TRITONSERVER_TYPE_INT16
public static final int TRITONSERVER_TYPE_INT32
public static final int TRITONSERVER_TYPE_INT64
public static final int TRITONSERVER_TYPE_FP16
public static final int TRITONSERVER_TYPE_FP32
public static final int TRITONSERVER_TYPE_FP64
public static final int TRITONSERVER_TYPE_BYTES
public static final int TRITONSERVER_TYPE_BF16
public static final int TRITONSERVER_MEMORY_CPU
public static final int TRITONSERVER_MEMORY_CPU_PINNED
public static final int TRITONSERVER_MEMORY_GPU
public static final int TRITONSERVER_PARAMETER_STRING
public static final int TRITONSERVER_PARAMETER_INT
public static final int TRITONSERVER_PARAMETER_BOOL
public static final int TRITONSERVER_PARAMETER_BYTES
public static final int TRITONSERVER_INSTANCEGROUPKIND_AUTO
public static final int TRITONSERVER_INSTANCEGROUPKIND_CPU
public static final int TRITONSERVER_INSTANCEGROUPKIND_GPU
public static final int TRITONSERVER_INSTANCEGROUPKIND_MODEL
public static final int TRITONSERVER_LOG_INFO
public static final int TRITONSERVER_LOG_WARN
public static final int TRITONSERVER_LOG_ERROR
public static final int TRITONSERVER_LOG_VERBOSE
public static final int TRITONSERVER_LOG_DEFAULT
public static final int TRITONSERVER_LOG_ISO8601
public static final int TRITONSERVER_ERROR_UNKNOWN
public static final int TRITONSERVER_ERROR_INTERNAL
public static final int TRITONSERVER_ERROR_NOT_FOUND
public static final int TRITONSERVER_ERROR_INVALID_ARG
public static final int TRITONSERVER_ERROR_UNAVAILABLE
public static final int TRITONSERVER_ERROR_UNSUPPORTED
public static final int TRITONSERVER_ERROR_ALREADY_EXISTS
public static final int TRITONSERVER_ERROR_CANCELLED
public static final int TRITONSERVER_METRIC_PROMETHEUS
public static final int TRITONSERVER_TRACE_LEVEL_DISABLED
public static final int TRITONSERVER_TRACE_LEVEL_MIN
public static final int TRITONSERVER_TRACE_LEVEL_MAX
public static final int TRITONSERVER_TRACE_LEVEL_TIMESTAMPS
public static final int TRITONSERVER_TRACE_LEVEL_TENSORS
public static final int TRITONSERVER_TRACE_REQUEST_START
public static final int TRITONSERVER_TRACE_QUEUE_START
public static final int TRITONSERVER_TRACE_COMPUTE_START
public static final int TRITONSERVER_TRACE_COMPUTE_INPUT_END
public static final int TRITONSERVER_TRACE_COMPUTE_OUTPUT_START
public static final int TRITONSERVER_TRACE_COMPUTE_END
public static final int TRITONSERVER_TRACE_REQUEST_END
public static final int TRITONSERVER_TRACE_TENSOR_QUEUE_INPUT
public static final int TRITONSERVER_TRACE_TENSOR_BACKEND_INPUT
public static final int TRITONSERVER_TRACE_TENSOR_BACKEND_OUTPUT
public static final int TRITONSERVER_REQUEST_FLAG_SEQUENCE_START
public static final int TRITONSERVER_REQUEST_FLAG_SEQUENCE_END
public static final int TRITONSERVER_REQUEST_RELEASE_ALL
public static final int TRITONSERVER_REQUEST_RELEASE_RESCHEDULE
public static final int TRITONSERVER_RESPONSE_COMPLETE_FINAL
public static final int TRITONSERVER_MODEL_CONTROL_NONE
public static final int TRITONSERVER_MODEL_CONTROL_POLL
public static final int TRITONSERVER_MODEL_CONTROL_EXPLICIT
public static final int TRITONSERVER_RATE_LIMIT_OFF
public static final int TRITONSERVER_RATE_LIMIT_EXEC_COUNT
public static final int TRITONSERVER_BATCH_UNKNOWN
public static final int TRITONSERVER_BATCH_FIRST_DIM
public static final int TRITONSERVER_INDEX_FLAG_READY
public static final int TRITONSERVER_TXN_ONE_TO_ONE
public static final int TRITONSERVER_TXN_DECOUPLED
public static final int TRITONSERVER_METRIC_KIND_COUNTER
public static final int TRITONSERVER_METRIC_KIND_GAUGE
public static final int TRITONBACKEND_API_VERSION_MAJOR
public static final int TRITONBACKEND_API_VERSION_MINOR
public static final int TRITONBACKEND_ARTIFACT_FILESYSTEM
public static final int TRITONBACKEND_EXECUTION_BLOCKING
public static final int TRITONBACKEND_EXECUTION_DEVICE_BLOCKING
public static final int TRITONREPOAGENT_API_VERSION_MAJOR
public static final int TRITONREPOAGENT_API_VERSION_MINOR
public static final int TRITONREPOAGENT_ARTIFACT_FILESYSTEM
public static final int TRITONREPOAGENT_ARTIFACT_REMOTE_FILESYSTEM
public static final int TRITONREPOAGENT_ACTION_LOAD
public static final int TRITONREPOAGENT_ACTION_LOAD_COMPLETE
public static final int TRITONREPOAGENT_ACTION_LOAD_FAIL
public static final int TRITONREPOAGENT_ACTION_UNLOAD
public static final int TRITONREPOAGENT_ACTION_UNLOAD_COMPLETE
public static TRITONSERVER_Error TRITONSERVER_ApiVersion(@Cast(value="uint32_t*") IntPointer major, @Cast(value="uint32_t*") IntPointer minor)
major
- Returns the TRITONSERVER API major version supported
by Triton.minor
- Returns the TRITONSERVER API minor version supported
by Triton.public static TRITONSERVER_Error TRITONSERVER_ApiVersion(@Cast(value="uint32_t*") IntBuffer major, @Cast(value="uint32_t*") IntBuffer minor)
public static TRITONSERVER_Error TRITONSERVER_ApiVersion(@Cast(value="uint32_t*") int[] major, @Cast(value="uint32_t*") int[] minor)
public static String TRITONSERVER_DataTypeString(@Cast(value="TRITONSERVER_DataType") int datatype)
datatype
- The data type.@Cast(value="TRITONSERVER_DataType") public static int TRITONSERVER_StringToDataType(String dtype)
dtype
- The datatype string representation.@Cast(value="TRITONSERVER_DataType") public static int TRITONSERVER_StringToDataType(@Cast(value="const char*") BytePointer dtype)
@Cast(value="uint32_t") public static int TRITONSERVER_DataTypeByteSize(@Cast(value="TRITONSERVER_DataType") int datatype)
dtype
- The datatype.public static String TRITONSERVER_MemoryTypeString(@Cast(value="TRITONSERVER_MemoryType") int memtype)
memtype
- The memory type.public static String TRITONSERVER_ParameterTypeString(@Cast(value="TRITONSERVER_ParameterType") int paramtype)
paramtype
- The parameter type.public static TRITONSERVER_Parameter TRITONSERVER_ParameterNew(String name, @Cast(value="const TRITONSERVER_ParameterType") int type, @Const Pointer value)
name
- The parameter name.type
- The parameter type.value
- The pointer to the value.public static TRITONSERVER_Parameter TRITONSERVER_ParameterNew(@Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_ParameterType") int type, @Const Pointer value)
public static TRITONSERVER_Parameter TRITONSERVER_ParameterBytesNew(String name, @Const Pointer byte_ptr, @Cast(value="const uint64_t") long size)
name
- The parameter name.byte_ptr
- The pointer to the data content.size
- The size of the data content.public static TRITONSERVER_Parameter TRITONSERVER_ParameterBytesNew(@Cast(value="const char*") BytePointer name, @Const Pointer byte_ptr, @Cast(value="const uint64_t") long size)
public static void TRITONSERVER_ParameterDelete(TRITONSERVER_Parameter parameter)
parameter
- The parameter object.public static String TRITONSERVER_InstanceGroupKindString(@Cast(value="TRITONSERVER_InstanceGroupKind") int kind)
kind
- The instance-group kind.@Cast(value="bool") public static boolean TRITONSERVER_LogIsEnabled(@Cast(value="TRITONSERVER_LogLevel") int level)
level
- The log level.public static TRITONSERVER_Error TRITONSERVER_LogMessage(@Cast(value="TRITONSERVER_LogLevel") int level, String filename, int line, String msg)
level
- The log level.filename
- The file name of the location of the log message.line
- The line number of the log message.msg
- The log message.public static TRITONSERVER_Error TRITONSERVER_LogMessage(@Cast(value="TRITONSERVER_LogLevel") int level, @Cast(value="const char*") BytePointer filename, int line, @Cast(value="const char*") BytePointer msg)
public static TRITONSERVER_Error TRITONSERVER_ErrorNew(@Cast(value="TRITONSERVER_Error_Code") int code, String msg)
code
- The error code.msg
- The error message.public static TRITONSERVER_Error TRITONSERVER_ErrorNew(@Cast(value="TRITONSERVER_Error_Code") int code, @Cast(value="const char*") BytePointer msg)
public static void TRITONSERVER_ErrorDelete(TRITONSERVER_Error error)
error
- The error object.@Cast(value="TRITONSERVER_Error_Code") public static int TRITONSERVER_ErrorCode(TRITONSERVER_Error error)
error
- The error object.public static String TRITONSERVER_ErrorCodeString(TRITONSERVER_Error error)
error
- The error object.public static String TRITONSERVER_ErrorMessage(TRITONSERVER_Error error)
error
- The error object.public static TRITONSERVER_Error TRITONSERVER_ResponseAllocatorNew(@Cast(value="TRITONSERVER_ResponseAllocator**") PointerPointer allocator, TRITONSERVER_ResponseAllocatorAllocFn_t alloc_fn, TRITONSERVER_ResponseAllocatorReleaseFn_t release_fn, TRITONSERVER_ResponseAllocatorStartFn_t start_fn)
allocator
- Returns the new response allocator object.alloc_fn
- The function to call to allocate buffers for result
tensors.release_fn
- The function to call when the server no longer
holds a reference to an allocated buffer.start_fn
- The function to call to indicate that the
subsequent 'alloc_fn' calls are for a new response. This callback
is optional (use nullptr to indicate that it should not be
invoked).public static TRITONSERVER_Error TRITONSERVER_ResponseAllocatorNew(@ByPtrPtr TRITONSERVER_ResponseAllocator allocator, TRITONSERVER_ResponseAllocatorAllocFn_t alloc_fn, TRITONSERVER_ResponseAllocatorReleaseFn_t release_fn, TRITONSERVER_ResponseAllocatorStartFn_t start_fn)
public static TRITONSERVER_Error TRITONSERVER_ResponseAllocatorSetBufferAttributesFunction(TRITONSERVER_ResponseAllocator allocator, TRITONSERVER_ResponseAllocatorBufferAttributesFn_t buffer_attributes_fn)
allocator
- The response allocator object.buffer_attributes_fn
- The function to call to get the buffer
attributes information for an allocated buffer.public static TRITONSERVER_Error TRITONSERVER_ResponseAllocatorSetQueryFunction(TRITONSERVER_ResponseAllocator allocator, TRITONSERVER_ResponseAllocatorQueryFn_t query_fn)
allocator
- The response allocator object.query_fn
- The function to call to query allocator's preferred memory
type and memory type ID.public static TRITONSERVER_Error TRITONSERVER_ResponseAllocatorDelete(TRITONSERVER_ResponseAllocator allocator)
allocator
- The response allocator object.public static TRITONSERVER_Error TRITONSERVER_MessageNewFromSerializedJson(@Cast(value="TRITONSERVER_Message**") PointerPointer message, String base, @Cast(value="size_t") long byte_size)
Create a new message object from serialized JSON string.
message
- The message object.base
- The base of the serialized JSON.byte_size
- The size, in bytes, of the serialized message.public static TRITONSERVER_Error TRITONSERVER_MessageNewFromSerializedJson(@ByPtrPtr TRITONSERVER_Message message, String base, @Cast(value="size_t") long byte_size)
public static TRITONSERVER_Error TRITONSERVER_MessageNewFromSerializedJson(@ByPtrPtr TRITONSERVER_Message message, @Cast(value="const char*") BytePointer base, @Cast(value="size_t") long byte_size)
public static TRITONSERVER_Error TRITONSERVER_MessageDelete(TRITONSERVER_Message message)
message
- The message object.public static TRITONSERVER_Error TRITONSERVER_MessageSerializeToJson(TRITONSERVER_Message message, @Cast(value="const char**") PointerPointer base, @Cast(value="size_t*") SizeTPointer byte_size)
message
- The message object.base
- Returns the base of the serialized message.byte_size
- Returns the size, in bytes, of the serialized
message.public static TRITONSERVER_Error TRITONSERVER_MessageSerializeToJson(TRITONSERVER_Message message, @Cast(value="const char**") @ByPtrPtr BytePointer base, @Cast(value="size_t*") SizeTPointer byte_size)
public static TRITONSERVER_Error TRITONSERVER_MessageSerializeToJson(TRITONSERVER_Message message, @Cast(value="const char**") @ByPtrPtr ByteBuffer base, @Cast(value="size_t*") SizeTPointer byte_size)
public static TRITONSERVER_Error TRITONSERVER_MessageSerializeToJson(TRITONSERVER_Message message, @Cast(value="const char**") @ByPtrPtr byte[] base, @Cast(value="size_t*") SizeTPointer byte_size)
public static TRITONSERVER_Error TRITONSERVER_MetricsDelete(TRITONSERVER_Metrics metrics)
metrics
- The metrics object.public static TRITONSERVER_Error TRITONSERVER_MetricsFormatted(TRITONSERVER_Metrics metrics, @Cast(value="TRITONSERVER_MetricFormat") int format, @Cast(value="const char**") PointerPointer base, @Cast(value="size_t*") SizeTPointer byte_size)
metrics
- The metrics object.format
- The format to use for the returned metrics.base
- Returns a pointer to the base of the formatted
metrics, as described above.byte_size
- Returns the size, in bytes, of the formatted
metrics.public static TRITONSERVER_Error TRITONSERVER_MetricsFormatted(TRITONSERVER_Metrics metrics, @Cast(value="TRITONSERVER_MetricFormat") int format, @Cast(value="const char**") @ByPtrPtr BytePointer base, @Cast(value="size_t*") SizeTPointer byte_size)
public static TRITONSERVER_Error TRITONSERVER_MetricsFormatted(TRITONSERVER_Metrics metrics, @Cast(value="TRITONSERVER_MetricFormat") int format, @Cast(value="const char**") @ByPtrPtr ByteBuffer base, @Cast(value="size_t*") SizeTPointer byte_size)
public static TRITONSERVER_Error TRITONSERVER_MetricsFormatted(TRITONSERVER_Metrics metrics, @Cast(value="TRITONSERVER_MetricFormat") int format, @Cast(value="const char**") @ByPtrPtr byte[] base, @Cast(value="size_t*") SizeTPointer byte_size)
public static String TRITONSERVER_InferenceTraceLevelString(@Cast(value="TRITONSERVER_InferenceTraceLevel") int level)
level
- The trace level.public static String TRITONSERVER_InferenceTraceActivityString(@Cast(value="TRITONSERVER_InferenceTraceActivity") int activity)
activity
- The trace activity.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceNew(@Cast(value="TRITONSERVER_InferenceTrace**") PointerPointer trace, @Cast(value="TRITONSERVER_InferenceTraceLevel") int level, @Cast(value="uint64_t") long parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, Pointer trace_userp)
trace
- Returns the new inference trace object.level
- The tracing level.parent_id
- The parent trace id for this trace. A value of 0
indicates that there is not parent trace.activity_fn
- The callback function where activity for the
trace is reported.release_fn
- The callback function called when all activity
is complete for the trace.trace_userp
- User-provided pointer that is delivered to
the activity and release callback functions.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceNew(@ByPtrPtr TRITONSERVER_InferenceTrace trace, @Cast(value="TRITONSERVER_InferenceTraceLevel") int level, @Cast(value="uint64_t") long parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, Pointer trace_userp)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceTensorNew(@Cast(value="TRITONSERVER_InferenceTrace**") PointerPointer trace, @Cast(value="TRITONSERVER_InferenceTraceLevel") int level, @Cast(value="uint64_t") long parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceTensorActivityFn_t tensor_activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, Pointer trace_userp)
trace
- Returns the new inference trace object.level
- The tracing level.parent_id
- The parent trace id for this trace. A value of 0
indicates that there is not parent trace.activity_fn
- The callback function where timeline activity for the
trace is reported.tensor_activity_fn
- The callback function where tensor activity for
the trace is reported.release_fn
- The callback function called when all activity
is complete for the trace.trace_userp
- User-provided pointer that is delivered to
the activity and release callback functions.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceTensorNew(@ByPtrPtr TRITONSERVER_InferenceTrace trace, @Cast(value="TRITONSERVER_InferenceTraceLevel") int level, @Cast(value="uint64_t") long parent_id, TRITONSERVER_InferenceTraceActivityFn_t activity_fn, TRITONSERVER_InferenceTraceTensorActivityFn_t tensor_activity_fn, TRITONSERVER_InferenceTraceReleaseFn_t release_fn, Pointer trace_userp)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceDelete(TRITONSERVER_InferenceTrace trace)
trace
- The trace object.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") LongPointer id)
trace
- The trace.id
- Returns the id associated with the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") LongBuffer id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") long[] id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceParentId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") LongPointer parent_id)
trace
- The trace.id
- Returns the parent id associated with the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceParentId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") LongBuffer parent_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceParentId(TRITONSERVER_InferenceTrace trace, @Cast(value="uint64_t*") long[] parent_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelName(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") PointerPointer model_name)
trace
- The trace.model_name
- Returns the name of the model associated with
the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelName(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr BytePointer model_name)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelName(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr ByteBuffer model_name)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelName(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr byte[] model_name)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelVersion(TRITONSERVER_InferenceTrace trace, @Cast(value="int64_t*") LongPointer model_version)
trace
- The trace.model_version
- Returns the version of the model associated
with the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelVersion(TRITONSERVER_InferenceTrace trace, @Cast(value="int64_t*") LongBuffer model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceModelVersion(TRITONSERVER_InferenceTrace trace, @Cast(value="int64_t*") long[] model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceRequestId(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") PointerPointer request_id)
trace
- The trace.request_id
- Returns the version of the model associated
with the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceRequestId(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr BytePointer request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceRequestId(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr ByteBuffer request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceRequestId(TRITONSERVER_InferenceTrace trace, @Cast(value="const char**") @ByPtrPtr byte[] request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceTraceSpawnChildTrace(TRITONSERVER_InferenceTrace trace, @Cast(value="TRITONSERVER_InferenceTrace**") PointerPointer child_trace)
trace
- The trace.child_trace
- Returns the child trace, spawned from the trace.public static TRITONSERVER_Error TRITONSERVER_InferenceTraceSpawnChildTrace(TRITONSERVER_InferenceTrace trace, @ByPtrPtr TRITONSERVER_InferenceTrace child_trace)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestNew(@Cast(value="TRITONSERVER_InferenceRequest**") PointerPointer inference_request, TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version)
inference_request
- Returns the new request object.server
- the inference server object.model_name
- The name of the model to use for the request.model_version
- The version of the model to use for the
request. If -1 then the server will choose a version based on the
model's policy.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestNew(@ByPtrPtr TRITONSERVER_InferenceRequest inference_request, TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestNew(@ByPtrPtr TRITONSERVER_InferenceRequest inference_request, TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestDelete(TRITONSERVER_InferenceRequest inference_request)
inference_request
- The request object.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") PointerPointer id)
inference_request
- The request object.id
- Returns the ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr BytePointer id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr ByteBuffer id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr byte[] id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetId(TRITONSERVER_InferenceRequest inference_request, String id)
inference_request
- The request object.id
- The ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestFlags(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") IntPointer flags)
inference_request
- The request object.flags
- Returns the flags.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestFlags(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") IntBuffer flags)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestFlags(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") int[] flags)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetFlags(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t") int flags)
inference_request
- The request object.flags
- The flags.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongPointer correlation_id)
inference_request
- The request object.correlation_id
- Returns the correlation ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongBuffer correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") long[] correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") PointerPointer correlation_id)
inference_request
- The request object.correlation_id
- Returns the correlation ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr BytePointer correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr ByteBuffer correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char**") @ByPtrPtr byte[] correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetCorrelationId(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t") long correlation_id)
inference_request
- The request object.correlation_id
- The correlation ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, String correlation_id)
inference_request
- The request object.correlation_id
- The correlation ID.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetCorrelationIdString(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer correlation_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestCancel(TRITONSERVER_InferenceRequest inference_request)
inference_request
- The request object.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestIsCancelled(TRITONSERVER_InferenceRequest inference_request, @Cast(value="bool*") boolean[] is_cancelled)
inference_request
- The request object.is_cancelled
- Returns whether the inference request is cancelled or
not.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestIsCancelled(TRITONSERVER_InferenceRequest inference_request, @Cast(value="bool*") BoolPointer is_cancelled)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriority(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") IntPointer priority)
inference_request
- The request object.priority
- Returns the priority level.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriority(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") IntBuffer priority)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriority(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t*") int[] priority)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriorityUInt64(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongPointer priority)
inference_request
- The request object.priority
- Returns the priority level.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriorityUInt64(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongBuffer priority)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestPriorityUInt64(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") long[] priority)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetPriority(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint32_t") int priority)
inference_request
- The request object.priority
- The priority level.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetPriorityUInt64(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t") long priority)
inference_request
- The request object.priority
- The priority level.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestTimeoutMicroseconds(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongPointer timeout_us)
inference_request
- The request object.timeout_us
- Returns the timeout, in microseconds.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestTimeoutMicroseconds(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") LongBuffer timeout_us)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestTimeoutMicroseconds(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t*") long[] timeout_us)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetTimeoutMicroseconds(TRITONSERVER_InferenceRequest inference_request, @Cast(value="uint64_t") long timeout_us)
inference_request
- The request object.timeout_us
- The timeout, in microseconds.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="uint64_t") long dim_count)
inference_request
- The request object.name
- The name of the input.datatype
- The type of the input. Valid type names are BOOL,
UINT8, UINT16, UINT32, UINT64, INT8, INT16, INT32, INT64, FP16,
FP32, FP64, and BYTES.shape
- The shape of the input.dim_count
- The number of dimensions of 'shape'.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="uint64_t") long dim_count)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="uint64_t") long dim_count)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="uint64_t") long dim_count)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="uint64_t") long dim_count)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddInput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="uint64_t") long dim_count)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddRawInput(TRITONSERVER_InferenceRequest inference_request, String name)
inference_request
- The request object.name
- The name of the input. This name is only used as a reference
of the raw input in other Tritonserver APIs. It doesn't associate with the
name used in the model.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddRawInput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveInput(TRITONSERVER_InferenceRequest inference_request, String name)
inference_request
- The request object.name
- The name of the input.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveInput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveAllInputs(TRITONSERVER_InferenceRequest inference_request)
inference_request
- The request object.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputData(TRITONSERVER_InferenceRequest inference_request, String name, @Const Pointer base, @Cast(value="size_t") long byte_size, @Cast(value="TRITONSERVER_MemoryType") int memory_type, @Cast(value="int64_t") long memory_type_id)
inference_request
- The request object.name
- The name of the input.base
- The base address of the input data.byte_size
- The size, in bytes, of the input data.memory_type
- The memory type of the input data.memory_type_id
- The memory type id of the input data.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputData(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Const Pointer base, @Cast(value="size_t") long byte_size, @Cast(value="TRITONSERVER_MemoryType") int memory_type, @Cast(value="int64_t") long memory_type_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputDataWithHostPolicy(TRITONSERVER_InferenceRequest inference_request, String name, @Const Pointer base, @Cast(value="size_t") long byte_size, @Cast(value="TRITONSERVER_MemoryType") int memory_type, @Cast(value="int64_t") long memory_type_id, String host_policy_name)
inference_request
- The request object.name
- The name of the input.base
- The base address of the input data.byte_size
- The size, in bytes, of the input data.memory_type
- The memory type of the input data.memory_type_id
- The memory type id of the input data.host_policy_name
- All model instances executing with this host_policy
will use this input buffer for execution.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputDataWithHostPolicy(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Const Pointer base, @Cast(value="size_t") long byte_size, @Cast(value="TRITONSERVER_MemoryType") int memory_type, @Cast(value="int64_t") long memory_type_id, @Cast(value="const char*") BytePointer host_policy_name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputDataWithBufferAttributes(TRITONSERVER_InferenceRequest inference_request, String name, @Const Pointer base, TRITONSERVER_BufferAttributes buffer_attributes)
inference_request
- The request object.name
- The name of the input.base
- The base address of the input data.buffer_attributes
- The buffer attributes of the input.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAppendInputDataWithBufferAttributes(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name, @Const Pointer base, TRITONSERVER_BufferAttributes buffer_attributes)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveAllInputData(TRITONSERVER_InferenceRequest inference_request, String name)
inference_request
- The request object.name
- The name of the input.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveAllInputData(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddRequestedOutput(TRITONSERVER_InferenceRequest inference_request, String name)
inference_request
- The request object.name
- The name of the output.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestAddRequestedOutput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveRequestedOutput(TRITONSERVER_InferenceRequest inference_request, String name)
inference_request
- The request object.name
- The name of the output.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveRequestedOutput(TRITONSERVER_InferenceRequest inference_request, @Cast(value="const char*") BytePointer name)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestRemoveAllRequestedOutputs(TRITONSERVER_InferenceRequest inference_request)
inference_request
- The request object.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetReleaseCallback(TRITONSERVER_InferenceRequest inference_request, TRITONSERVER_InferenceRequestReleaseFn_t request_release_fn, Pointer request_release_userp)
inference_request
- The request object.request_release_fn
- The function called to return ownership
of the 'inference_request' object.request_release_userp
- User-provided pointer that is
delivered to the 'request_release_fn' callback.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetResponseCallback(TRITONSERVER_InferenceRequest inference_request, TRITONSERVER_ResponseAllocator response_allocator, Pointer response_allocator_userp, TRITONSERVER_InferenceResponseCompleteFn_t response_fn, Pointer response_userp)
inference_request
- The request object.response_allocator
- The TRITONSERVER_ResponseAllocator to use
to allocate buffers to hold inference results.response_allocator_userp
- User-provided pointer that is
delivered to the response allocator's start and allocation functions.response_fn
- The function called to deliver an inference
response for this request.response_userp
- User-provided pointer that is delivered to
the 'response_fn' callback.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetStringParameter(TRITONSERVER_InferenceRequest request, String key, String value)
request
- The request.key
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetStringParameter(TRITONSERVER_InferenceRequest request, @Cast(value="const char*") BytePointer key, @Cast(value="const char*") BytePointer value)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetIntParameter(TRITONSERVER_InferenceRequest request, String key, @Cast(value="const int64_t") long value)
request
- The request.key
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetIntParameter(TRITONSERVER_InferenceRequest request, @Cast(value="const char*") BytePointer key, @Cast(value="const int64_t") long value)
public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetBoolParameter(TRITONSERVER_InferenceRequest request, String key, @Cast(value="const bool") boolean value)
request
- The request.key
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONSERVER_InferenceRequestSetBoolParameter(TRITONSERVER_InferenceRequest request, @Cast(value="const char*") BytePointer key, @Cast(value="const bool") boolean value)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseDelete(TRITONSERVER_InferenceResponse inference_response)
Delete an inference response object.
inference_response
- The response object.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseError(TRITONSERVER_InferenceResponse inference_response)
inference_response
- The response object.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseModel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") PointerPointer model_name, @Cast(value="int64_t*") LongPointer model_version)
inference_response
- The response object.model_name
- Returns the name of the model.model_version
- Returns the version of the model.
this response.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseModel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr BytePointer model_name, @Cast(value="int64_t*") LongPointer model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseModel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr ByteBuffer model_name, @Cast(value="int64_t*") LongBuffer model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseModel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr byte[] model_name, @Cast(value="int64_t*") long[] model_version)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseId(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") PointerPointer request_id)
inference_response
- The response object.request_id
- Returns the ID of the request corresponding to
this response.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseId(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr BytePointer request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseId(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr ByteBuffer request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseId(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const char**") @ByPtrPtr byte[] request_id)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameterCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") IntPointer count)
inference_response
- The response object.count
- Returns the number of parameters.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameterCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameterCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameter(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer name, @Cast(value="TRITONSERVER_ParameterType*") IntPointer type, @Cast(value="const void**") PointerPointer vvalue)
inference_response
- The response object.index
- The index of the parameter, must be 0 <= index <
count, where 'count' is the value returned by
TRITONSERVER_InferenceResponseParameterCount.name
- Returns the name of the parameter.type
- Returns the type of the parameter.vvalue
- Returns a pointer to the parameter value.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameter(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_ParameterType*") IntPointer type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameter(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_ParameterType*") IntBuffer type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseParameter(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_ParameterType*") int[] type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") IntPointer count)
inference_response
- The response object.count
- Returns the number of output tensors.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputCount(TRITONSERVER_InferenceResponse inference_response, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutput(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") PointerPointer shape, @Cast(value="uint64_t*") LongPointer dim_count, @Cast(value="const void**") PointerPointer base, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id, @Cast(value="void**") PointerPointer userp)
inference_response
- The response object.index
- The index of the output tensor, must be 0 <= index <
count, where 'count' is the value returned by
TRITONSERVER_InferenceResponseOutputCount.name
- Returns the name of the output.datatype
- Returns the type of the output.shape
- Returns the shape of the output.dim_count
- Returns the number of dimensions of the returned
shape.base
- Returns the tensor data for the output.byte_size
- Returns the size, in bytes, of the data.memory_type
- Returns the memory type of the data.memory_type_id
- Returns the memory type id of the data.userp
- The user-specified value associated with the buffer
in TRITONSERVER_ResponseAllocatorAllocFn_t.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutput(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint64_t*") LongPointer dim_count, @Cast(value="const void**") @ByPtrPtr Pointer base, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id, @Cast(value="void**") @ByPtrPtr Pointer userp)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutput(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint64_t*") LongBuffer dim_count, @Cast(value="const void**") @ByPtrPtr Pointer base, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id, @Cast(value="void**") @ByPtrPtr Pointer userp)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutput(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint64_t*") long[] dim_count, @Cast(value="const void**") @ByPtrPtr Pointer base, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id, @Cast(value="void**") @ByPtrPtr Pointer userp)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputClassificationLabel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const size_t") long class_index, @Cast(value="const char**") PointerPointer label)
inference_response
- The response object.index
- The index of the output tensor, must be 0 <= index <
count, where 'count' is the value returned by
TRITONSERVER_InferenceResponseOutputCount.class_index
- The index of the class.name
- Returns the label corresponding to 'class_index' or
nullptr if no label.public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputClassificationLabel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const size_t") long class_index, @Cast(value="const char**") @ByPtrPtr BytePointer label)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputClassificationLabel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const size_t") long class_index, @Cast(value="const char**") @ByPtrPtr ByteBuffer label)
public static TRITONSERVER_Error TRITONSERVER_InferenceResponseOutputClassificationLabel(TRITONSERVER_InferenceResponse inference_response, @Cast(value="const uint32_t") int index, @Cast(value="const size_t") long class_index, @Cast(value="const char**") @ByPtrPtr byte[] label)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesNew(@Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer buffer_attributes)
Create a new buffer attributes object. The caller takes ownership of the TRITONSERVER_BufferAttributes object and must call TRITONSERVER_BufferAttributesDelete to release the object.
buffer_attributes
- Returns the new buffer attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesNew(@ByPtrPtr TRITONSERVER_BufferAttributes buffer_attributes)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesDelete(TRITONSERVER_BufferAttributes buffer_attributes)
buffer_attributes
- The buffer_attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesSetMemoryTypeId(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="int64_t") long memory_type_id)
buffer_attributes
- The buffer attributes object.memory_type_id
- Memory type id to assign to the buffer attributes
object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesSetMemoryType(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="TRITONSERVER_MemoryType") int memory_type)
buffer_attributes
- The buffer attributes object.memory_type
- Memory type to assign to the buffer attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesSetCudaIpcHandle(TRITONSERVER_BufferAttributes buffer_attributes, Pointer cuda_ipc_handle)
buffer_attributes
- The buffer attributes object.cuda_ipc_handle
- The CudaIpcHandle to assign to the buffer attributes
object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesSetByteSize(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="size_t") long byte_size)
buffer_attributes
- The buffer attributes object.byte_size
- Byte size to assign to the buffer attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryTypeId(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="int64_t*") LongPointer memory_type_id)
buffer_attributes
- The buffer attributes object.memory_type_id
- Returns the memory type id associated with the buffer
attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryTypeId(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryTypeId(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryType(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type)
buffer_attributes
- The buffer attributes object.memory_type
- Returns the memory type associated with the buffer
attributes object.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryType(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesMemoryType(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesCudaIpcHandle(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="void**") PointerPointer cuda_ipc_handle)
buffer_attributes
- The buffer attributes object.cuda_ipc_handle
- Returns the memory type associated with the buffer
attributes object. If the cudaIpcHandle does not exist for the buffer,
nullptr will be returned.public static TRITONSERVER_Error TRITONSERVER_BufferAttributesCudaIpcHandle(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="void**") @ByPtrPtr Pointer cuda_ipc_handle)
public static TRITONSERVER_Error TRITONSERVER_BufferAttributesByteSize(TRITONSERVER_BufferAttributes buffer_attributes, @Cast(value="size_t*") SizeTPointer byte_size)
buffer_attributes
- The buffer attributes object.byte_size
- Returns the byte size associated with the buffer attributes
object.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsNew(@Cast(value="TRITONSERVER_ServerOptions**") PointerPointer options)
options
- Returns the new server options object.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsNew(@ByPtrPtr TRITONSERVER_ServerOptions options)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsDelete(TRITONSERVER_ServerOptions options)
options
- The server options object.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetServerId(TRITONSERVER_ServerOptions options, String server_id)
options
- The server options object.server_id
- The server identifier.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetServerId(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer server_id)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelRepositoryPath(TRITONSERVER_ServerOptions options, String model_repository_path)
options
- The server options object.model_repository_path
- The full path to the model repository.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelRepositoryPath(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer model_repository_path)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelControlMode(TRITONSERVER_ServerOptions options, @Cast(value="TRITONSERVER_ModelControlMode") int mode)
options
- The server options object.mode
- The mode to use for the model control.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetStartupModel(TRITONSERVER_ServerOptions options, String model_name)
options
- The server options object.mode_name
- The name of the model to load on startup.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetStartupModel(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer model_name)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetStrictModelConfig(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean strict)
options
- The server options object.strict
- True to enable strict model configuration handling,
false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetRateLimiterMode(TRITONSERVER_ServerOptions options, @Cast(value="TRITONSERVER_RateLimitMode") int mode)
options
- The server options object.mode
- The mode to use for the rate limiting. By default, execution
count is used to determine the priorities.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsAddRateLimiterResource(TRITONSERVER_ServerOptions options, String resource_name, @Cast(value="const size_t") long resource_count, int device)
options
- The server options object.name
- The name of the resource.count
- The count of the resource.device
- The device identifier for the resource. A value of -1
indicates that the specified number of resources are available on every
device. The device value is ignored for a global resource. The server
will use the rate limiter configuration specified for instance groups
in model config to determine whether resource is global. In case of
conflicting resource type in different model configurations, server
will raise an appropriate error while loading model.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsAddRateLimiterResource(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer resource_name, @Cast(value="const size_t") long resource_count, int device)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetPinnedMemoryPoolByteSize(TRITONSERVER_ServerOptions options, @Cast(value="uint64_t") long size)
options
- The server options object.size
- The pinned memory pool byte size.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCudaMemoryPoolByteSize(TRITONSERVER_ServerOptions options, int gpu_device, @Cast(value="uint64_t") long size)
options
- The server options object.gpu_device
- The GPU device to allocate the memory pool.size
- The CUDA memory pool byte size.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCudaVirtualAddressSize(TRITONSERVER_ServerOptions options, int gpu_device, @Cast(value="size_t") long cuda_virtual_address_size)
options
- The server options object.gpu_device
- The GPU device to set the CUDA virtual address space sizesize
- The size of the CUDA virtual address space.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetResponseCacheByteSize(TRITONSERVER_ServerOptions options, @Cast(value="uint64_t") long size)
options
- The server options object.size
- The total response cache byte size.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCacheConfig(TRITONSERVER_ServerOptions options, String cache_name, String config_json)
options
- The server options object.cache_name
- The name of the cache. Example names would be
"local", "redis", or the name of a custom cache implementation.config_json
- The string representation of config JSON that is
used to initialize the cache implementation.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCacheConfig(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer cache_name, @Cast(value="const char*") BytePointer config_json)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCacheDirectory(TRITONSERVER_ServerOptions options, String cache_dir)
options
- The server options object.cache_dir
- The full path of the cache directory.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCacheDirectory(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer cache_dir)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetMinSupportedComputeCapability(TRITONSERVER_ServerOptions options, double cc)
options
- The server options object.cc
- The minimum CUDA compute capability.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetExitOnError(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean exit)
options
- The server options object.exit
- True to enable exiting on initialization error, false
to continue.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetStrictReadiness(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean strict)
options
- The server options object.strict
- True to enable strict readiness handling, false to
disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetExitTimeout(TRITONSERVER_ServerOptions options, @Cast(value="unsigned int") int timeout)
options
- The server options object.timeout
- The exit timeout, in seconds.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetBufferManagerThreadCount(TRITONSERVER_ServerOptions options, @Cast(value="unsigned int") int thread_count)
options
- The server options object.thread_count
- The number of threads.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelLoadThreadCount(TRITONSERVER_ServerOptions options, @Cast(value="unsigned int") int thread_count)
options
- The server options object.thread_count
- The number of threads.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelNamespacing(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean enable_namespace)
options
- The server options object.enable_namespace
- Whether to enable model namespacing or not.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogFile(TRITONSERVER_ServerOptions options, String file)
options
- The server options object.file
- a string defining the file where the log outputs will be saved.
An empty string for the file name will cause triton to direct logging
facilities to the consolepublic static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogFile(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer file)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogInfo(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean log)
options
- The server options object.log
- True to enable info logging, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogWarn(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean log)
options
- The server options object.log
- True to enable warning logging, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogError(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean log)
options
- The server options object.log
- True to enable error logging, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogFormat(TRITONSERVER_ServerOptions options, @Cast(value="const TRITONSERVER_LogFormat") int format)
options
- The server options object.format
- The logging format.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetLogVerbose(TRITONSERVER_ServerOptions options, int level)
options
- The server options object.level
- The verbose logging level.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetMetrics(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean metrics)
options
- The server options object.metrics
- True to enable metrics, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetGpuMetrics(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean gpu_metrics)
options
- The server options object.gpu_metrics
- True to enable GPU metrics, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetCpuMetrics(TRITONSERVER_ServerOptions options, @Cast(value="bool") boolean cpu_metrics)
options
- The server options object.cpu_metrics
- True to enable CPU metrics, false to disable.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetMetricsInterval(TRITONSERVER_ServerOptions options, @Cast(value="uint64_t") long metrics_interval_ms)
options
- The server options object.metrics_interval_ms
- The time interval in ms between
successive metrics updates.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetBackendDirectory(TRITONSERVER_ServerOptions options, String backend_dir)
options
- The server options object.backend_dir
- The full path of the backend directory.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetBackendDirectory(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer backend_dir)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetRepoAgentDirectory(TRITONSERVER_ServerOptions options, String repoagent_dir)
options
- The server options object.repoagent_dir
- The full path of the repository agent directory.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetRepoAgentDirectory(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer repoagent_dir)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetModelLoadDeviceLimit(TRITONSERVER_ServerOptions options, @Cast(value="const TRITONSERVER_InstanceGroupKind") int kind, int device_id, double fraction)
options
- The server options object.kind
- The kind of the device.device_id
- The id of the device.fraction
- The limit on memory usage as a fractionpublic static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetBackendConfig(TRITONSERVER_ServerOptions options, String backend_name, String setting, String value)
options
- The server options object.backend_name
- The name of the backend.setting
- The name of the setting.value
- The setting value.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetBackendConfig(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer backend_name, @Cast(value="const char*") BytePointer setting, @Cast(value="const char*") BytePointer value)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetHostPolicy(TRITONSERVER_ServerOptions options, String policy_name, String setting, String value)
options
- The server options object.policy_name
- The name of the policy.setting
- The name of the setting.value
- The setting value.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetHostPolicy(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer policy_name, @Cast(value="const char*") BytePointer setting, @Cast(value="const char*") BytePointer value)
public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetMetricsConfig(TRITONSERVER_ServerOptions options, String name, String setting, String value)
options
- The server options object.name
- The name of the configuration group. An empty string indicates
a global configuration option.setting
- The name of the setting.value
- The setting value.public static TRITONSERVER_Error TRITONSERVER_ServerOptionsSetMetricsConfig(TRITONSERVER_ServerOptions options, @Cast(value="const char*") BytePointer name, @Cast(value="const char*") BytePointer setting, @Cast(value="const char*") BytePointer value)
public static TRITONSERVER_Error TRITONSERVER_ServerNew(@Cast(value="TRITONSERVER_Server**") PointerPointer server, TRITONSERVER_ServerOptions options)
server
- Returns the new inference server object.options
- The inference server options object.public static TRITONSERVER_Error TRITONSERVER_ServerNew(@ByPtrPtr TRITONSERVER_Server server, TRITONSERVER_ServerOptions options)
public static TRITONSERVER_Error TRITONSERVER_ServerDelete(TRITONSERVER_Server server)
server
- The inference server object.public static TRITONSERVER_Error TRITONSERVER_ServerStop(TRITONSERVER_Server server)
server
- The inference server object.public static TRITONSERVER_Error TRITONSERVER_ServerRegisterModelRepository(TRITONSERVER_Server server, String repository_path, @Cast(value="const TRITONSERVER_Parameter**") PointerPointer name_mapping, @Cast(value="const uint32_t") int mapping_count)
server
- The inference server object.repository_path
- The full path to the model repository.name_mapping
- List of name_mapping parameters. Each mapping has
the model directory name as its key, overridden model name as its value.model_count
- Number of mappings provided.public static TRITONSERVER_Error TRITONSERVER_ServerRegisterModelRepository(TRITONSERVER_Server server, String repository_path, @Const @ByPtrPtr TRITONSERVER_Parameter name_mapping, @Cast(value="const uint32_t") int mapping_count)
public static TRITONSERVER_Error TRITONSERVER_ServerRegisterModelRepository(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer repository_path, @Const @ByPtrPtr TRITONSERVER_Parameter name_mapping, @Cast(value="const uint32_t") int mapping_count)
public static TRITONSERVER_Error TRITONSERVER_ServerUnregisterModelRepository(TRITONSERVER_Server server, String repository_path)
server
- The inference server object.repository_path
- The full path to the model repository.public static TRITONSERVER_Error TRITONSERVER_ServerUnregisterModelRepository(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer repository_path)
public static TRITONSERVER_Error TRITONSERVER_ServerPollModelRepository(TRITONSERVER_Server server)
server
- The inference server object.public static TRITONSERVER_Error TRITONSERVER_ServerIsLive(TRITONSERVER_Server server, @Cast(value="bool*") boolean[] live)
server
- The inference server object.live
- Returns true if server is live, false otherwise.public static TRITONSERVER_Error TRITONSERVER_ServerIsLive(TRITONSERVER_Server server, @Cast(value="bool*") BoolPointer live)
public static TRITONSERVER_Error TRITONSERVER_ServerIsReady(TRITONSERVER_Server server, @Cast(value="bool*") boolean[] ready)
server
- The inference server object.ready
- Returns true if server is ready, false otherwise.public static TRITONSERVER_Error TRITONSERVER_ServerIsReady(TRITONSERVER_Server server, @Cast(value="bool*") BoolPointer ready)
public static TRITONSERVER_Error TRITONSERVER_ServerModelIsReady(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="bool*") boolean[] ready)
server
- The inference server object.model_name
- The name of the model to get readiness for.model_version
- The version of the model to get readiness
for. If -1 then the server will choose a version based on the
model's policy.ready
- Returns true if server is ready, false otherwise.public static TRITONSERVER_Error TRITONSERVER_ServerModelIsReady(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="bool*") BoolPointer ready)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer flags, @Cast(value="void**") PointerPointer voidp)
server
- The inference server object.model_name
- The name of the model.model_version
- The version of the model. If -1 then the
server will choose a version based on the model's policy.flags
- Returns flags indicating the batch properties of the
model.voidp
- If non-nullptr, returns a point specific to the
'flags' value.public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntBuffer flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") int[] flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntBuffer flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelBatchProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") int[] flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer txn_flags, @Cast(value="void**") PointerPointer voidp)
server
- The inference server object.model_name
- The name of the model.model_version
- The version of the model. If -1 then the
server will choose a version based on the model's policy.txn_flags
- Returns flags indicating the transaction policy of the
model.voidp
- If non-nullptr, returns a point specific to the 'flags' value.public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntBuffer txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") int[] txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntPointer txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") IntBuffer txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerModelTransactionProperties(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="uint32_t*") int[] txn_flags, @Cast(value="void**") @ByPtrPtr Pointer voidp)
public static TRITONSERVER_Error TRITONSERVER_ServerMetadata(TRITONSERVER_Server server, @Cast(value="TRITONSERVER_Message**") PointerPointer server_metadata)
server
- The inference server object.server_metadata
- Returns the server metadata message.public static TRITONSERVER_Error TRITONSERVER_ServerMetadata(TRITONSERVER_Server server, @ByPtrPtr TRITONSERVER_Message server_metadata)
public static TRITONSERVER_Error TRITONSERVER_ServerModelMetadata(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="TRITONSERVER_Message**") PointerPointer model_metadata)
server
- The inference server object.model_name
- The name of the model.model_version
- The version of the model.
If -1 then the server will choose a version based on the model's
policy.model_metadata
- Returns the model metadata message.public static TRITONSERVER_Error TRITONSERVER_ServerModelMetadata(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @ByPtrPtr TRITONSERVER_Message model_metadata)
public static TRITONSERVER_Error TRITONSERVER_ServerModelMetadata(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @ByPtrPtr TRITONSERVER_Message model_metadata)
public static TRITONSERVER_Error TRITONSERVER_ServerModelStatistics(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="TRITONSERVER_Message**") PointerPointer model_stats)
server
- The inference server object.model_name
- The name of the model.
If empty, then statistics for all available models will be returned,
and the server will choose a version based on those models' policies.model_version
- The version of the model. If -1 then the
server will choose a version based on the model's policy.model_stats
- Returns the model statistics message.public static TRITONSERVER_Error TRITONSERVER_ServerModelStatistics(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @ByPtrPtr TRITONSERVER_Message model_stats)
public static TRITONSERVER_Error TRITONSERVER_ServerModelStatistics(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @ByPtrPtr TRITONSERVER_Message model_stats)
public static TRITONSERVER_Error TRITONSERVER_ServerModelConfig(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="const uint32_t") int config_version, @Cast(value="TRITONSERVER_Message**") PointerPointer model_config)
server
- The inference server object.model_name
- The name of the model.model_version
- The version of the model. If -1 then the
server will choose a version based on the model's policy.config_version
- The model configuration will be returned in
a format matching this version. If the configuration cannot be
represented in the requested version's format then an error will
be returned. Currently only version 1 is supported.model_config
- Returns the model config message.public static TRITONSERVER_Error TRITONSERVER_ServerModelConfig(TRITONSERVER_Server server, String model_name, @Cast(value="const int64_t") long model_version, @Cast(value="const uint32_t") int config_version, @ByPtrPtr TRITONSERVER_Message model_config)
public static TRITONSERVER_Error TRITONSERVER_ServerModelConfig(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Cast(value="const int64_t") long model_version, @Cast(value="const uint32_t") int config_version, @ByPtrPtr TRITONSERVER_Message model_config)
public static TRITONSERVER_Error TRITONSERVER_ServerModelIndex(TRITONSERVER_Server server, @Cast(value="uint32_t") int flags, @Cast(value="TRITONSERVER_Message**") PointerPointer model_index)
server
- The inference server object.flags
- TRITONSERVER_ModelIndexFlag flags that control how to
collect the index.model_index
- Return the model index message that holds the
index of all models contained in the server's model repository(s).public static TRITONSERVER_Error TRITONSERVER_ServerModelIndex(TRITONSERVER_Server server, @Cast(value="uint32_t") int flags, @ByPtrPtr TRITONSERVER_Message model_index)
public static TRITONSERVER_Error TRITONSERVER_ServerLoadModel(TRITONSERVER_Server server, String model_name)
server
- The inference server object.model_name
- The name of the model.public static TRITONSERVER_Error TRITONSERVER_ServerLoadModel(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name)
public static TRITONSERVER_Error TRITONSERVER_ServerLoadModelWithParameters(TRITONSERVER_Server server, String model_name, @Cast(value="const TRITONSERVER_Parameter**") PointerPointer parameters, @Cast(value="const uint64_t") long parameter_count)
server
- The inference server object.model_name
- The name of the model.parameters
- The array of load parameters.parameter_count
- The number of parameters.public static TRITONSERVER_Error TRITONSERVER_ServerLoadModelWithParameters(TRITONSERVER_Server server, String model_name, @Const @ByPtrPtr TRITONSERVER_Parameter parameters, @Cast(value="const uint64_t") long parameter_count)
public static TRITONSERVER_Error TRITONSERVER_ServerLoadModelWithParameters(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name, @Const @ByPtrPtr TRITONSERVER_Parameter parameters, @Cast(value="const uint64_t") long parameter_count)
public static TRITONSERVER_Error TRITONSERVER_ServerUnloadModel(TRITONSERVER_Server server, String model_name)
server
- The inference server object.model_name
- The name of the model.public static TRITONSERVER_Error TRITONSERVER_ServerUnloadModel(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name)
public static TRITONSERVER_Error TRITONSERVER_ServerUnloadModelAndDependents(TRITONSERVER_Server server, String model_name)
server
- The inference server object.model_name
- The name of the model.public static TRITONSERVER_Error TRITONSERVER_ServerUnloadModelAndDependents(TRITONSERVER_Server server, @Cast(value="const char*") BytePointer model_name)
public static TRITONSERVER_Error TRITONSERVER_ServerMetrics(TRITONSERVER_Server server, @Cast(value="TRITONSERVER_Metrics**") PointerPointer metrics)
server
- The inference server object.metrics
- Returns the metrics.public static TRITONSERVER_Error TRITONSERVER_ServerMetrics(TRITONSERVER_Server server, @ByPtrPtr TRITONSERVER_Metrics metrics)
public static TRITONSERVER_Error TRITONSERVER_ServerInferAsync(TRITONSERVER_Server server, TRITONSERVER_InferenceRequest inference_request, TRITONSERVER_InferenceTrace trace)
server
- The inference server object.inference_request
- The request object.trace
- The trace object for this request, or nullptr if no
tracing.public static TRITONSERVER_Error TRITONSERVER_MetricFamilyNew(@Cast(value="TRITONSERVER_MetricFamily**") PointerPointer family, @Cast(value="const TRITONSERVER_MetricKind") int kind, String name, String description)
family
- Returns the new metric family object.kind
- The type of metric family to create.name
- The name of the metric family seen when calling the metrics
endpoint.description
- The description of the metric family seen when
calling the metrics endpoint.public static TRITONSERVER_Error TRITONSERVER_MetricFamilyNew(@ByPtrPtr TRITONSERVER_MetricFamily family, @Cast(value="const TRITONSERVER_MetricKind") int kind, String name, String description)
public static TRITONSERVER_Error TRITONSERVER_MetricFamilyNew(@ByPtrPtr TRITONSERVER_MetricFamily family, @Cast(value="const TRITONSERVER_MetricKind") int kind, @Cast(value="const char*") BytePointer name, @Cast(value="const char*") BytePointer description)
public static TRITONSERVER_Error TRITONSERVER_MetricFamilyDelete(TRITONSERVER_MetricFamily family)
family
- The metric family object to delete.public static TRITONSERVER_Error TRITONSERVER_MetricNew(@Cast(value="TRITONSERVER_Metric**") PointerPointer metric, TRITONSERVER_MetricFamily family, @Cast(value="const TRITONSERVER_Parameter**") PointerPointer labels, @Cast(value="const uint64_t") long label_count)
metric
- Returns the new metric object.family
- The metric family to add this new metric to.labels
- The array of labels to associate with this new metric.label_count
- The number of labels.public static TRITONSERVER_Error TRITONSERVER_MetricNew(@ByPtrPtr TRITONSERVER_Metric metric, TRITONSERVER_MetricFamily family, @Const @ByPtrPtr TRITONSERVER_Parameter labels, @Cast(value="const uint64_t") long label_count)
public static TRITONSERVER_Error TRITONSERVER_MetricDelete(TRITONSERVER_Metric metric)
metric
- The metric object to delete.public static TRITONSERVER_Error TRITONSERVER_MetricValue(TRITONSERVER_Metric metric, DoublePointer value)
metric
- The metric object to query.value
- Returns the current value of the metric object.public static TRITONSERVER_Error TRITONSERVER_MetricValue(TRITONSERVER_Metric metric, DoubleBuffer value)
public static TRITONSERVER_Error TRITONSERVER_MetricValue(TRITONSERVER_Metric metric, double[] value)
public static TRITONSERVER_Error TRITONSERVER_MetricIncrement(TRITONSERVER_Metric metric, double value)
metric
- The metric object to update.value
- The amount to increment the metric's value by.public static TRITONSERVER_Error TRITONSERVER_MetricSet(TRITONSERVER_Metric metric, double value)
metric
- The metric object to update.value
- The amount to set metric's value to.public static TRITONSERVER_Error TRITONSERVER_GetMetricKind(TRITONSERVER_Metric metric, @Cast(value="TRITONSERVER_MetricKind*") IntPointer kind)
metric
- The metric object to query.kind
- Returns the TRITONSERVER_MetricKind of metric.public static TRITONSERVER_Error TRITONSERVER_GetMetricKind(TRITONSERVER_Metric metric, @Cast(value="TRITONSERVER_MetricKind*") IntBuffer kind)
public static TRITONSERVER_Error TRITONSERVER_GetMetricKind(TRITONSERVER_Metric metric, @Cast(value="TRITONSERVER_MetricKind*") int[] kind)
public static TRITONSERVER_Error TRITONBACKEND_ApiVersion(@Cast(value="uint32_t*") IntPointer major, @Cast(value="uint32_t*") IntPointer minor)
major
- Returns the TRITONBACKEND API major version supported
by Triton.minor
- Returns the TRITONBACKEND API minor version supported
by Triton.public static TRITONSERVER_Error TRITONBACKEND_ApiVersion(@Cast(value="uint32_t*") IntBuffer major, @Cast(value="uint32_t*") IntBuffer minor)
public static TRITONSERVER_Error TRITONBACKEND_ApiVersion(@Cast(value="uint32_t*") int[] major, @Cast(value="uint32_t*") int[] minor)
public static TRITONSERVER_Error TRITONBACKEND_MemoryManagerAllocate(TRITONBACKEND_MemoryManager manager, @Cast(value="void**") PointerPointer buffer, @Cast(value="const TRITONSERVER_MemoryType") int memory_type, @Cast(value="const int64_t") long memory_type_id, @Cast(value="const uint64_t") long byte_size)
Allocate a contiguous block of memory of a specific type using a memory manager. Two error codes have specific interpretations for this function: TRITONSERVER_ERROR_UNSUPPORTED: Indicates that Triton is incapable of allocating the requested memory type and memory type ID. Requests for the memory type and ID will always fail no matter 'byte_size' of the request. TRITONSERVER_ERROR_UNAVAILABLE: Indicates that Triton can allocate the memory type and ID but that currently it cannot allocate a contiguous block of memory of the requested 'byte_size'.
manager
- The memory manager.buffer
- Returns the allocated memory.memory_type
- The type of memory to allocate.memory_type_id
- The ID associated with the memory type to
allocate. For GPU memory this indicates the device ID of the GPU
to allocate from.byte_size
- The size of memory to allocate, in bytes.public static TRITONSERVER_Error TRITONBACKEND_MemoryManagerAllocate(TRITONBACKEND_MemoryManager manager, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const TRITONSERVER_MemoryType") int memory_type, @Cast(value="const int64_t") long memory_type_id, @Cast(value="const uint64_t") long byte_size)
public static TRITONSERVER_Error TRITONBACKEND_MemoryManagerFree(TRITONBACKEND_MemoryManager manager, Pointer buffer, @Cast(value="const TRITONSERVER_MemoryType") int memory_type, @Cast(value="const int64_t") long memory_type_id)
manager
- The memory manager.buffer
- The allocated memory buffer to free.memory_type
- The type of memory of the buffer.memory_type_id
- The ID associated with the memory type of
the buffer.public static TRITONSERVER_Error TRITONBACKEND_InputProperties(TRITONBACKEND_Input input, @Cast(value="const char**") PointerPointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") PointerPointer shape, @Cast(value="uint32_t*") IntPointer dims_count, @Cast(value="uint64_t*") LongPointer byte_size, @Cast(value="uint32_t*") IntPointer buffer_count)
Get the name and properties of an input tensor. The returned strings and other properties are owned by the input, not the caller, and so should not be modified or freed.
input
- The input tensor.name
- If non-nullptr, returns the tensor name.datatype
- If non-nullptr, returns the tensor datatype.shape
- If non-nullptr, returns the tensor shape.dim_count
- If non-nullptr, returns the number of dimensions
in the tensor shape.byte_size
- If non-nullptr, returns the size of the available
data for the tensor, in bytes. This size reflects the actual data
available, and does not necessarily match what is
expected/required for the tensor given its shape and datatype. It
is the responsibility of the backend to handle mismatches in these
sizes appropriately.buffer_count
- If non-nullptr, returns the number of buffers
holding the contents of the tensor. These buffers are accessed
using TRITONBACKEND_InputBuffer.public static TRITONSERVER_Error TRITONBACKEND_InputProperties(TRITONBACKEND_Input input, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint32_t*") IntPointer dims_count, @Cast(value="uint64_t*") LongPointer byte_size, @Cast(value="uint32_t*") IntPointer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputProperties(TRITONBACKEND_Input input, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint32_t*") IntBuffer dims_count, @Cast(value="uint64_t*") LongBuffer byte_size, @Cast(value="uint32_t*") IntBuffer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputProperties(TRITONBACKEND_Input input, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint32_t*") int[] dims_count, @Cast(value="uint64_t*") long[] byte_size, @Cast(value="uint32_t*") int[] buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const char**") PointerPointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") PointerPointer shape, @Cast(value="uint32_t*") IntPointer dims_count, @Cast(value="uint64_t*") LongPointer byte_size, @Cast(value="uint32_t*") IntPointer buffer_count)
input
- The input tensor.host_policy_name
- The host policy name. Fallback input properties
will be return if nullptr is provided.name
- If non-nullptr, returns the tensor name.datatype
- If non-nullptr, returns the tensor datatype.shape
- If non-nullptr, returns the tensor shape.dim_count
- If non-nullptr, returns the number of dimensions
in the tensor shape.byte_size
- If non-nullptr, returns the size of the available
data for the tensor, in bytes. This size reflects the actual data
available, and does not necessarily match what is
expected/required for the tensor given its shape and datatype. It
is the responsibility of the backend to handle mismatches in these
sizes appropriately.buffer_count
- If non-nullptr, returns the number of buffers
holding the contents of the tensor. These buffers are accessed
using TRITONBACKEND_InputBufferForHostPolicy.public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint32_t*") IntPointer dims_count, @Cast(value="uint64_t*") LongPointer byte_size, @Cast(value="uint32_t*") IntPointer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint32_t*") IntBuffer dims_count, @Cast(value="uint64_t*") LongBuffer byte_size, @Cast(value="uint32_t*") IntBuffer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint32_t*") int[] dims_count, @Cast(value="uint64_t*") long[] byte_size, @Cast(value="uint32_t*") int[] buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint32_t*") IntPointer dims_count, @Cast(value="uint64_t*") LongPointer byte_size, @Cast(value="uint32_t*") IntPointer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint32_t*") IntBuffer dims_count, @Cast(value="uint64_t*") LongBuffer byte_size, @Cast(value="uint32_t*") IntBuffer buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputPropertiesForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint32_t*") int[] dims_count, @Cast(value="uint64_t*") long[] byte_size, @Cast(value="uint32_t*") int[] buffer_count)
public static TRITONSERVER_Error TRITONBACKEND_InputBuffer(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") PointerPointer buffer, @Cast(value="uint64_t*") LongPointer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
input
- The input tensor.index
- The index of the buffer. Must be 0 <= index <
buffer_count, where buffer_count is the value returned by
TRITONBACKEND_InputProperties.buffer
- Returns a pointer to a contiguous block of data for
the named input.buffer_byte_size
- Returns the size, in bytes, of 'buffer'.memory_type
- Acts as both input and output. On input gives
the buffer memory type preferred by the function caller. Returns
the actual memory type of 'buffer'.memory_type_id
- Acts as both input and output. On input
gives the buffer memory type id preferred by the function caller.
Returns the actual memory type id of 'buffer'.public static TRITONSERVER_Error TRITONBACKEND_InputBuffer(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongPointer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBuffer(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongBuffer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBuffer(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") long[] buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") PointerPointer buffer, @Cast(value="uint64_t*") LongPointer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
input
- The input tensor.host_policy_name
- The host policy name. Fallback input buffer
will be return if nullptr is provided.index
- The index of the buffer. Must be 0 <= index <
buffer_count, where buffer_count is the value returned by
TRITONBACKEND_InputPropertiesForHostPolicy.buffer
- Returns a pointer to a contiguous block of data for
the named input.buffer_byte_size
- Returns the size, in bytes, of 'buffer'.memory_type
- Acts as both input and output. On input gives
the buffer memory type preferred by the function caller. Returns
the actual memory type of 'buffer'.memory_type_id
- Acts as both input and output. On input
gives the buffer memory type id preferred by the function caller.
Returns the actual memory type id of 'buffer'.public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongPointer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongBuffer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") long[] buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongPointer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, String host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") LongBuffer buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferForHostPolicy(TRITONBACKEND_Input input, @Cast(value="const char*") BytePointer host_policy_name, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @Cast(value="uint64_t*") long[] buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_InputBufferAttributes(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") PointerPointer buffer, @Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer buffer_attributes)
input
- The input tensor.index
- The index of the buffer. Must be 0 <= index < buffer_count,
where buffer_count is the value returned by TRITONBACKEND_InputProperties.buffer
- Returns a pointer to a contiguous block of data for
the named input.buffer_attributes
- Returns the attributes for the given buffer.public static TRITONSERVER_Error TRITONBACKEND_InputBufferAttributes(TRITONBACKEND_Input input, @Cast(value="const uint32_t") int index, @Cast(value="const void**") @ByPtrPtr Pointer buffer, @ByPtrPtr TRITONSERVER_BufferAttributes buffer_attributes)
public static TRITONSERVER_Error TRITONBACKEND_OutputBuffer(TRITONBACKEND_Output output, @Cast(value="void**") PointerPointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
Get a buffer to use to hold the tensor data for the output. The returned buffer is owned by the output and so should not be freed by the caller. The caller can and should fill the buffer with the output data for the tensor. The lifetime of the buffer matches that of the output and so the buffer should not be accessed after the output tensor object is released.
buffer
- Returns a pointer to a buffer where the contents of
the output tensor should be placed.buffer_byte_size
- The size, in bytes, of the buffer required
by the caller.memory_type
- Acts as both input and output. On input gives
the buffer memory type preferred by the caller. Returns the
actual memory type of 'buffer'.memory_type_id
- Acts as both input and output. On input
gives the buffer memory type id preferred by the caller. Returns
the actual memory type id of 'buffer'.public static TRITONSERVER_Error TRITONBACKEND_OutputBuffer(TRITONBACKEND_Output output, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_OutputBuffer(TRITONBACKEND_Output output, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_OutputBuffer(TRITONBACKEND_Output output, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_OutputBufferAttributes(TRITONBACKEND_Output output, @Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer buffer_attributes)
output
- The output tensor.buffer_attributes
- Returns the attributes for the output buffer.public static TRITONSERVER_Error TRITONBACKEND_OutputBufferAttributes(TRITONBACKEND_Output output, @ByPtrPtr TRITONSERVER_BufferAttributes buffer_attributes)
public static TRITONSERVER_Error TRITONBACKEND_RequestId(TRITONBACKEND_Request request, @Cast(value="const char**") PointerPointer id)
Get the ID of the request. Can be nullptr if request doesn't have an ID. The returned string is owned by the request, not the caller, and so should not be modified or freed.
request
- The inference request.id
- Returns the ID.public static TRITONSERVER_Error TRITONBACKEND_RequestId(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr BytePointer id)
public static TRITONSERVER_Error TRITONBACKEND_RequestId(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr ByteBuffer id)
public static TRITONSERVER_Error TRITONBACKEND_RequestId(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr byte[] id)
public static TRITONSERVER_Error TRITONBACKEND_RequestIsCancelled(TRITONBACKEND_Request request, @Cast(value="bool*") boolean[] is_cancelled)
request
- The inference request.is_cancelled
- Returns true if the request is cancelled otherwise it
would return false.public static TRITONSERVER_Error TRITONBACKEND_RequestIsCancelled(TRITONBACKEND_Request request, @Cast(value="bool*") BoolPointer is_cancelled)
public static TRITONSERVER_Error TRITONBACKEND_ResponseFactoryIsCancelled(TRITONBACKEND_ResponseFactory factory, @Cast(value="bool*") boolean[] is_cancelled)
factory
- The response factoryis_cancelled
- Returns true if the request is cancelled otherwise it
would return false.public static TRITONSERVER_Error TRITONBACKEND_ResponseFactoryIsCancelled(TRITONBACKEND_ResponseFactory factory, @Cast(value="bool*") BoolPointer is_cancelled)
public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationId(TRITONBACKEND_Request request, @Cast(value="uint64_t*") LongPointer id)
request
- The inference request.id
- Returns the correlation ID.public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationId(TRITONBACKEND_Request request, @Cast(value="uint64_t*") LongBuffer id)
public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationId(TRITONBACKEND_Request request, @Cast(value="uint64_t*") long[] id)
public static TRITONSERVER_Error TRITONBACKEND_InferenceRequestTimeoutMicroseconds(TRITONBACKEND_Request request, @Cast(value="uint64_t*") LongPointer timeout)
request
- The inference request.timeout
- Returns the timeout value for the request.public static TRITONSERVER_Error TRITONBACKEND_InferenceRequestTimeoutMicroseconds(TRITONBACKEND_Request request, @Cast(value="uint64_t*") LongBuffer timeout)
public static TRITONSERVER_Error TRITONBACKEND_InferenceRequestTimeoutMicroseconds(TRITONBACKEND_Request request, @Cast(value="uint64_t*") long[] timeout)
public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationIdString(TRITONBACKEND_Request request, @Cast(value="const char**") PointerPointer id)
request
- The inference request.id
- Returns the correlation ID.public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationIdString(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr BytePointer id)
public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationIdString(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr ByteBuffer id)
public static TRITONSERVER_Error TRITONBACKEND_RequestCorrelationIdString(TRITONBACKEND_Request request, @Cast(value="const char**") @ByPtrPtr byte[] id)
public static TRITONSERVER_Error TRITONBACKEND_RequestFlags(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntPointer flags)
request
- The inference request.flags
- Returns the flags.public static TRITONSERVER_Error TRITONBACKEND_RequestFlags(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntBuffer flags)
public static TRITONSERVER_Error TRITONBACKEND_RequestFlags(TRITONBACKEND_Request request, @Cast(value="uint32_t*") int[] flags)
public static TRITONSERVER_Error TRITONBACKEND_RequestParameterCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntPointer count)
request
- The inference request.count
- Returns the number of parameters.public static TRITONSERVER_Error TRITONBACKEND_RequestParameterCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONBACKEND_RequestParameterCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONBACKEND_RequestParameter(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer key, @Cast(value="TRITONSERVER_ParameterType*") IntPointer type, @Cast(value="const void**") PointerPointer vvalue)
request
- The inference request.index
- The index of the parameter. Must be 0 <= index <
count, where count is the value returned by
TRITONBACKEND_RequestParameterCount.key
- Returns the key of the parameter.type
- Returns the type of the parameter.vvalue
- Returns a pointer to the parameter value.public static TRITONSERVER_Error TRITONBACKEND_RequestParameter(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer key, @Cast(value="TRITONSERVER_ParameterType*") IntPointer type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONBACKEND_RequestParameter(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer key, @Cast(value="TRITONSERVER_ParameterType*") IntBuffer type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONBACKEND_RequestParameter(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] key, @Cast(value="TRITONSERVER_ParameterType*") int[] type, @Cast(value="const void**") @ByPtrPtr Pointer vvalue)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntPointer count)
request
- The inference request.count
- Returns the number of input tensors.public static TRITONSERVER_Error TRITONBACKEND_RequestInputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer input_name)
request
- The inference request.index
- The index of the input tensor. Must be 0 <= index <
count, where count is the value returned by
TRITONBACKEND_RequestInputCount.input_name
- Returns the name of the input tensor
corresponding to the index.public static TRITONSERVER_Error TRITONBACKEND_RequestInputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer input_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer input_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] input_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestInput(TRITONBACKEND_Request request, String name, @Cast(value="TRITONBACKEND_Input**") PointerPointer input)
request
- The inference request.name
- The name of the input.input
- Returns the input corresponding to the name.public static TRITONSERVER_Error TRITONBACKEND_RequestInput(TRITONBACKEND_Request request, String name, @ByPtrPtr TRITONBACKEND_Input input)
public static TRITONSERVER_Error TRITONBACKEND_RequestInput(TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @ByPtrPtr TRITONBACKEND_Input input)
public static TRITONSERVER_Error TRITONBACKEND_RequestInputByIndex(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="TRITONBACKEND_Input**") PointerPointer input)
request
- The inference request.index
- The index of the input tensor. Must be 0 <= index <
count, where count is the value returned by
TRITONBACKEND_RequestInputCount.input
- Returns the input corresponding to the index.public static TRITONSERVER_Error TRITONBACKEND_RequestInputByIndex(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @ByPtrPtr TRITONBACKEND_Input input)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntPointer count)
request
- The inference request.count
- Returns the number of output tensors.public static TRITONSERVER_Error TRITONBACKEND_RequestOutputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputCount(TRITONBACKEND_Request request, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer output_name)
request
- The inference request.index
- The index of the requested output tensor. Must be 0
<= index < count, where count is the value returned by
TRITONBACKEND_RequestOutputCount.output_name
- Returns the name of the requested output tensor
corresponding to the index.public static TRITONSERVER_Error TRITONBACKEND_RequestOutputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer output_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer output_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputName(TRITONBACKEND_Request request, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] output_name)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, String name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
request
- The request.name
- The name of the output tensor. This is optional
and it should be set to nullptr to indicate that the tensor name has
not determined.byte_size
- The expected size of the buffer. This is optional
and it should be set to nullptr to indicate that the byte size has
not determined.memory_type
- Acts as both input and output. On input gives
the memory type preferred by the caller. Returns memory type preferred
by Triton, taken account of the caller preferred type.memory_type_id
- Acts as both input and output. On input gives
the memory type ID preferred by the caller. Returns memory type ID preferred
by Triton, taken account of the caller preferred type ID.public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, String name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, String name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_RequestOutputBufferProperties(TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="size_t*") SizeTPointer byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_RequestRelease(TRITONBACKEND_Request request, @Cast(value="uint32_t") int release_flags)
request
- The inference request.release_flags
- Flags indicating what type of request release
should be performed. @see TRITONSERVER_RequestReleaseFlag. @see
TRITONSERVER_InferenceRequestReleaseFn_t.public static TRITONSERVER_Error TRITONBACKEND_RequestTrace(TRITONBACKEND_Request request, @Cast(value="TRITONSERVER_InferenceTrace**") PointerPointer trace)
nullptr
will be returned.request
- The inference request.trace
- Returns the trace associated with the request.public static TRITONSERVER_Error TRITONBACKEND_RequestTrace(TRITONBACKEND_Request request, @ByPtrPtr TRITONSERVER_InferenceTrace trace)
public static TRITONSERVER_Error TRITONBACKEND_ResponseFactoryNew(@Cast(value="TRITONBACKEND_ResponseFactory**") PointerPointer factory, TRITONBACKEND_Request request)
Create the response factory associated with a request.
factory
- Returns the new response factory.request
- The inference request.public static TRITONSERVER_Error TRITONBACKEND_ResponseFactoryNew(@ByPtrPtr TRITONBACKEND_ResponseFactory factory, TRITONBACKEND_Request request)
public static TRITONSERVER_Error TRITONBACKEND_ResponseFactoryDelete(TRITONBACKEND_ResponseFactory factory)
factory
- The response factory.public static TRITONSERVER_Error TRITONBACKEND_ResponseFactorySendFlags(TRITONBACKEND_ResponseFactory factory, @Cast(value="const uint32_t") int send_flags)
factory
- The response factory.send_flags
- Flags to send. @see
TRITONSERVER_ResponseCompleteFlag. @see
TRITONSERVER_InferenceResponseCompleteFn_t.public static TRITONSERVER_Error TRITONBACKEND_ResponseNew(@Cast(value="TRITONBACKEND_Response**") PointerPointer response, TRITONBACKEND_Request request)
Create a response for a request.
response
- Returns the new response.request
- The request.public static TRITONSERVER_Error TRITONBACKEND_ResponseNew(@ByPtrPtr TRITONBACKEND_Response response, TRITONBACKEND_Request request)
public static TRITONSERVER_Error TRITONBACKEND_ResponseNewFromFactory(@Cast(value="TRITONBACKEND_Response**") PointerPointer response, TRITONBACKEND_ResponseFactory factory)
response
- Returns the new response.factory
- The response factory.public static TRITONSERVER_Error TRITONBACKEND_ResponseNewFromFactory(@ByPtrPtr TRITONBACKEND_Response response, TRITONBACKEND_ResponseFactory factory)
public static TRITONSERVER_Error TRITONBACKEND_ResponseDelete(TRITONBACKEND_Response response)
response
- The response.public static TRITONSERVER_Error TRITONBACKEND_ResponseSetStringParameter(TRITONBACKEND_Response response, String name, String value)
response
- The response.name
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONBACKEND_ResponseSetStringParameter(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="const char*") BytePointer value)
public static TRITONSERVER_Error TRITONBACKEND_ResponseSetIntParameter(TRITONBACKEND_Response response, String name, @Cast(value="const int64_t") long value)
response
- The response.name
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONBACKEND_ResponseSetIntParameter(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="const int64_t") long value)
public static TRITONSERVER_Error TRITONBACKEND_ResponseSetBoolParameter(TRITONBACKEND_Response response, String name, @Cast(value="const bool") boolean value)
response
- The response.name
- The name of the parameter.value
- The value of the parameter.public static TRITONSERVER_Error TRITONBACKEND_ResponseSetBoolParameter(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="const bool") boolean value)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @Cast(value="TRITONBACKEND_Output**") PointerPointer output, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
response
- The response.output
- Returns the new response output.name
- The name of the output tensor.datatype
- The datatype of the output tensor.shape
- The shape of the output tensor.dims_count
- The number of dimensions in the output tensor
shape.public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseOutput(TRITONBACKEND_Response response, @ByPtrPtr TRITONBACKEND_Output output, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_ResponseSend(TRITONBACKEND_Response response, @Cast(value="const uint32_t") int send_flags, TRITONSERVER_Error error)
response
- The response.send_flags
- Flags associated with the response. @see
TRITONSERVER_ResponseCompleteFlag. @see
TRITONSERVER_InferenceResponseCompleteFn_t.error
- The TRITONSERVER_Error to send if the response is an
error, or nullptr if the response is successful.public static TRITONSERVER_Error TRITONBACKEND_StateNew(@Cast(value="TRITONBACKEND_State**") PointerPointer state, TRITONBACKEND_Request request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
Create a state in the request. The returned state object is only valid before the TRITONBACKEND_StateUpdate is called. The state should not be freed by the caller. If TRITONBACKEND_StateUpdate is not called, the lifetime of the state matches the lifetime of the request. If the state name does not exist in the "state" section of the model configuration, the state will not be created and an error will be returned. If this function is called when sequence batching is not enabled or there is no 'states' section in the sequence batching section of the model configuration, this call will return an error.
state
- Returns the new state.request
- The request.name
- The name of the state.datatype
- The datatype of the state.shape
- The shape of the state.dims_count
- The number of dimensions in the state shape.public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongPointer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, String name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") LongBuffer shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateNew(@ByPtrPtr TRITONBACKEND_State state, TRITONBACKEND_Request request, @Cast(value="const char*") BytePointer name, @Cast(value="const TRITONSERVER_DataType") int datatype, @Cast(value="const int64_t*") long[] shape, @Cast(value="const uint32_t") int dims_count)
public static TRITONSERVER_Error TRITONBACKEND_StateUpdate(TRITONBACKEND_State state)
state
- The state.public static TRITONSERVER_Error TRITONBACKEND_StateBuffer(TRITONBACKEND_State state, @Cast(value="void**") PointerPointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
state
- The state.buffer
- Returns a pointer to a buffer where the contents of the state
should be placed.buffer_byte_size
- The size, in bytes, of the buffer required
by the caller.memory_type
- Acts as both input and output. On input gives
the buffer memory type preferred by the caller. Returns the
actual memory type of 'buffer'.memory_type_id
- Acts as both input and output. On input
gives the buffer memory type id preferred by the caller. Returns
the actual memory type id of 'buffer'.public static TRITONSERVER_Error TRITONBACKEND_StateBuffer(TRITONBACKEND_State state, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntPointer memory_type, @Cast(value="int64_t*") LongPointer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_StateBuffer(TRITONBACKEND_State state, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") IntBuffer memory_type, @Cast(value="int64_t*") LongBuffer memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_StateBuffer(TRITONBACKEND_State state, @Cast(value="void**") @ByPtrPtr Pointer buffer, @Cast(value="const uint64_t") long buffer_byte_size, @Cast(value="TRITONSERVER_MemoryType*") int[] memory_type, @Cast(value="int64_t*") long[] memory_type_id)
public static TRITONSERVER_Error TRITONBACKEND_StateBufferAttributes(TRITONBACKEND_State state, @Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer buffer_attributes)
state
- The state.buffer_attributes
- Returns the buffer attributes for the given state.public static TRITONSERVER_Error TRITONBACKEND_StateBufferAttributes(TRITONBACKEND_State state, @ByPtrPtr TRITONSERVER_BufferAttributes buffer_attributes)
public static TRITONSERVER_Error TRITONBACKEND_BackendName(TRITONBACKEND_Backend backend, @Cast(value="const char**") PointerPointer name)
backend
- The backend.name
- Returns the name of the backend.public static TRITONSERVER_Error TRITONBACKEND_BackendName(TRITONBACKEND_Backend backend, @Cast(value="const char**") @ByPtrPtr BytePointer name)
public static TRITONSERVER_Error TRITONBACKEND_BackendName(TRITONBACKEND_Backend backend, @Cast(value="const char**") @ByPtrPtr ByteBuffer name)
public static TRITONSERVER_Error TRITONBACKEND_BackendName(TRITONBACKEND_Backend backend, @Cast(value="const char**") @ByPtrPtr byte[] name)
public static TRITONSERVER_Error TRITONBACKEND_BackendConfig(TRITONBACKEND_Backend backend, @Cast(value="TRITONSERVER_Message**") PointerPointer backend_config)
backend
- The backend.backend_config
- Returns the backend configuration as a message.public static TRITONSERVER_Error TRITONBACKEND_BackendConfig(TRITONBACKEND_Backend backend, @ByPtrPtr TRITONSERVER_Message backend_config)
public static TRITONSERVER_Error TRITONBACKEND_BackendExecutionPolicy(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ExecutionPolicy*") IntPointer policy)
backend
- The backend.policy
- Returns the execution policy.public static TRITONSERVER_Error TRITONBACKEND_BackendExecutionPolicy(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ExecutionPolicy*") IntBuffer policy)
public static TRITONSERVER_Error TRITONBACKEND_BackendExecutionPolicy(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ExecutionPolicy*") int[] policy)
public static TRITONSERVER_Error TRITONBACKEND_BackendSetExecutionPolicy(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ExecutionPolicy") int policy)
backend
- The backend.policy
- The execution policy.public static TRITONSERVER_Error TRITONBACKEND_BackendArtifacts(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") PointerPointer location)
backend
- The backend.artifact_type
- Returns the artifact type for the backend.path
- Returns the location.public static TRITONSERVER_Error TRITONBACKEND_BackendArtifacts(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") @ByPtrPtr BytePointer location)
public static TRITONSERVER_Error TRITONBACKEND_BackendArtifacts(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ArtifactType*") IntBuffer artifact_type, @Cast(value="const char**") @ByPtrPtr ByteBuffer location)
public static TRITONSERVER_Error TRITONBACKEND_BackendArtifacts(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_ArtifactType*") int[] artifact_type, @Cast(value="const char**") @ByPtrPtr byte[] location)
public static TRITONSERVER_Error TRITONBACKEND_BackendMemoryManager(TRITONBACKEND_Backend backend, @Cast(value="TRITONBACKEND_MemoryManager**") PointerPointer manager)
backend
- The backend.manager
- Returns the memory manager.public static TRITONSERVER_Error TRITONBACKEND_BackendMemoryManager(TRITONBACKEND_Backend backend, @ByPtrPtr TRITONBACKEND_MemoryManager manager)
public static TRITONSERVER_Error TRITONBACKEND_BackendState(TRITONBACKEND_Backend backend, @Cast(value="void**") PointerPointer state)
backend
- The backend.state
- Returns the user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_BackendState(TRITONBACKEND_Backend backend, @Cast(value="void**") @ByPtrPtr Pointer state)
public static TRITONSERVER_Error TRITONBACKEND_BackendSetState(TRITONBACKEND_Backend backend, Pointer state)
backend
- The backend.state
- The user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_ModelName(TRITONBACKEND_Model model, @Cast(value="const char**") PointerPointer name)
Get the name of the model. The returned string is owned by the model object, not the caller, and so should not be modified or freed.
model
- The model.name
- Returns the model name.public static TRITONSERVER_Error TRITONBACKEND_ModelName(TRITONBACKEND_Model model, @Cast(value="const char**") @ByPtrPtr BytePointer name)
public static TRITONSERVER_Error TRITONBACKEND_ModelName(TRITONBACKEND_Model model, @Cast(value="const char**") @ByPtrPtr ByteBuffer name)
public static TRITONSERVER_Error TRITONBACKEND_ModelName(TRITONBACKEND_Model model, @Cast(value="const char**") @ByPtrPtr byte[] name)
public static TRITONSERVER_Error TRITONBACKEND_ModelVersion(TRITONBACKEND_Model model, @Cast(value="uint64_t*") LongPointer version)
model
- The model.version
- Returns the model version.public static TRITONSERVER_Error TRITONBACKEND_ModelVersion(TRITONBACKEND_Model model, @Cast(value="uint64_t*") LongBuffer version)
public static TRITONSERVER_Error TRITONBACKEND_ModelVersion(TRITONBACKEND_Model model, @Cast(value="uint64_t*") long[] version)
public static TRITONSERVER_Error TRITONBACKEND_ModelRepository(TRITONBACKEND_Model model, @Cast(value="TRITONBACKEND_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") PointerPointer location)
model
- The model.artifact_type
- Returns the artifact type for the model.path
- Returns the location.public static TRITONSERVER_Error TRITONBACKEND_ModelRepository(TRITONBACKEND_Model model, @Cast(value="TRITONBACKEND_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") @ByPtrPtr BytePointer location)
public static TRITONSERVER_Error TRITONBACKEND_ModelRepository(TRITONBACKEND_Model model, @Cast(value="TRITONBACKEND_ArtifactType*") IntBuffer artifact_type, @Cast(value="const char**") @ByPtrPtr ByteBuffer location)
public static TRITONSERVER_Error TRITONBACKEND_ModelRepository(TRITONBACKEND_Model model, @Cast(value="TRITONBACKEND_ArtifactType*") int[] artifact_type, @Cast(value="const char**") @ByPtrPtr byte[] location)
public static TRITONSERVER_Error TRITONBACKEND_ModelConfig(TRITONBACKEND_Model model, @Cast(value="const uint32_t") int config_version, @Cast(value="TRITONSERVER_Message**") PointerPointer model_config)
model
- The model.config_version
- The model configuration will be returned in
a format matching this version. If the configuration cannot be
represented in the requested version's format then an error will
be returned. Currently only version 1 is supported.model_config
- Returns the model configuration as a message.public static TRITONSERVER_Error TRITONBACKEND_ModelConfig(TRITONBACKEND_Model model, @Cast(value="const uint32_t") int config_version, @ByPtrPtr TRITONSERVER_Message model_config)
public static TRITONSERVER_Error TRITONBACKEND_ModelAutoCompleteConfig(TRITONBACKEND_Model model, @Cast(value="bool*") boolean[] auto_complete_config)
model
- The model.auto_complete_config
- Returns whether the backend should auto-complete
the model configuration.public static TRITONSERVER_Error TRITONBACKEND_ModelAutoCompleteConfig(TRITONBACKEND_Model model, @Cast(value="bool*") BoolPointer auto_complete_config)
public static TRITONSERVER_Error TRITONBACKEND_ModelSetConfig(TRITONBACKEND_Model model, @Cast(value="const uint32_t") int config_version, TRITONSERVER_Message model_config)
model
- The model.config_version
- The format version of the model configuration.
If the configuration is not represented in the version's format
then an error will be returned. Currently only version 1 is supported.model_config
- The updated model configuration as a message.public static TRITONSERVER_Error TRITONBACKEND_ModelServer(TRITONBACKEND_Model model, @Cast(value="TRITONSERVER_Server**") PointerPointer server)
model
- The model.server
- Returns the server.public static TRITONSERVER_Error TRITONBACKEND_ModelServer(TRITONBACKEND_Model model, @ByPtrPtr TRITONSERVER_Server server)
public static TRITONSERVER_Error TRITONBACKEND_ModelBackend(TRITONBACKEND_Model model, @Cast(value="TRITONBACKEND_Backend**") PointerPointer backend)
model
- The model.model
- Returns the backend object.public static TRITONSERVER_Error TRITONBACKEND_ModelBackend(TRITONBACKEND_Model model, @ByPtrPtr TRITONBACKEND_Backend backend)
public static TRITONSERVER_Error TRITONBACKEND_ModelState(TRITONBACKEND_Model model, @Cast(value="void**") PointerPointer state)
model
- The model.state
- Returns the user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_ModelState(TRITONBACKEND_Model model, @Cast(value="void**") @ByPtrPtr Pointer state)
public static TRITONSERVER_Error TRITONBACKEND_ModelSetState(TRITONBACKEND_Model model, Pointer state)
model
- The model.state
- The user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_ModelReportMemoryUsage(TRITONBACKEND_Model model, @Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer usage, @Cast(value="uint32_t") int usage_size)
model
- The model.usage
- The list of buffer attributes that records the memory usage,
each entry should record the total memory usage of a given memory type and
id. For example, if the model itself occupies 64 bytes on each of
CUDA device 0 and CUDA device 1. Then 'usage' should have first two entries
set, one has the buffer attributes of "type GPU, id 0, 64 bytes" and the
other has "type GPU, id 1, 64 bytes". 'usage' is owned by the backend and
may be released after the function returns.usage_size
- The number of entries in 'usage'.public static TRITONSERVER_Error TRITONBACKEND_ModelReportMemoryUsage(TRITONBACKEND_Model model, @ByPtrPtr TRITONSERVER_BufferAttributes usage, @Cast(value="uint32_t") int usage_size)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceName(TRITONBACKEND_ModelInstance instance, @Cast(value="const char**") PointerPointer name)
Get the name of the model instance. The returned string is owned by the model object, not the caller, and so should not be modified or freed.
instance
- The model instance.name
- Returns the instance name.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceName(TRITONBACKEND_ModelInstance instance, @Cast(value="const char**") @ByPtrPtr BytePointer name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceName(TRITONBACKEND_ModelInstance instance, @Cast(value="const char**") @ByPtrPtr ByteBuffer name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceName(TRITONBACKEND_ModelInstance instance, @Cast(value="const char**") @ByPtrPtr byte[] name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceKind(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONSERVER_InstanceGroupKind*") IntPointer kind)
instance
- The model instance.kind
- Returns the instance kind.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceKind(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONSERVER_InstanceGroupKind*") IntBuffer kind)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceKind(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONSERVER_InstanceGroupKind*") int[] kind)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceDeviceId(TRITONBACKEND_ModelInstance instance, IntPointer device_id)
instance
- The model instance.device_id
- Returns the instance device ID.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceDeviceId(TRITONBACKEND_ModelInstance instance, IntBuffer device_id)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceDeviceId(TRITONBACKEND_ModelInstance instance, int[] device_id)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceHostPolicy(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONSERVER_Message**") PointerPointer host_policy)
instance
- The model instance.host_policy
- Returns the host policy setting as a message.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceHostPolicy(TRITONBACKEND_ModelInstance instance, @ByPtrPtr TRITONSERVER_Message host_policy)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceIsPassive(TRITONBACKEND_ModelInstance instance, @Cast(value="bool*") boolean[] is_passive)
instance
- The model instance.is_passive
- Returns true if the instance is passive, false otherwisepublic static TRITONSERVER_Error TRITONBACKEND_ModelInstanceIsPassive(TRITONBACKEND_ModelInstance instance, @Cast(value="bool*") BoolPointer is_passive)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") IntPointer count)
instance
- The model instance.count
- Returns the number of optimization profiles.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileName(TRITONBACKEND_ModelInstance instance, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer profile_name)
instance
- The model instance.index
- The index of the optimization profile. Must be 0
<= index < count, where count is the value returned by
TRITONBACKEND_ModelInstanceProfileCount.profile_name
- Returns the name of the optimization profile
corresponding to the index.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileName(TRITONBACKEND_ModelInstance instance, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer profile_name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileName(TRITONBACKEND_ModelInstance instance, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer profile_name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceProfileName(TRITONBACKEND_ModelInstance instance, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] profile_name)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") IntPointer count)
instance
- The model instance.count
- Returns the number of secondary devices.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceCount(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceProperties(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t") int index, @Cast(value="const char**") PointerPointer kind, @Cast(value="int64_t*") LongPointer id)
instance
- The model instance.index
- The index of the secondary device. Must be 0
<= index < count, where count is the value returned by
TRITONBACKEND_ModelInstanceSecondaryDeviceCount.kind
- Returns the kind of secondary device corresponding
to the index.id
- Returns the id of secondary device corresponding to the index.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceProperties(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer kind, @Cast(value="int64_t*") LongPointer id)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceProperties(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer kind, @Cast(value="int64_t*") LongBuffer id)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSecondaryDeviceProperties(TRITONBACKEND_ModelInstance instance, @Cast(value="uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] kind, @Cast(value="int64_t*") long[] id)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceModel(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONBACKEND_Model**") PointerPointer model)
instance
- The model instance.backend
- Returns the model object.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceModel(TRITONBACKEND_ModelInstance instance, @ByPtrPtr TRITONBACKEND_Model model)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceState(TRITONBACKEND_ModelInstance instance, @Cast(value="void**") PointerPointer state)
instance
- The model instance.state
- Returns the user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceState(TRITONBACKEND_ModelInstance instance, @Cast(value="void**") @ByPtrPtr Pointer state)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceSetState(TRITONBACKEND_ModelInstance instance, Pointer state)
instance
- The model instance.state
- The user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceReportMemoryUsage(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONSERVER_BufferAttributes**") PointerPointer usage, @Cast(value="uint32_t") int usage_size)
instance
- The model instance.usage
- The list of buffer attributes that records the memory usage,
each entry should record the total memory usage of a given memory type and
id. For example, if the instance itself occupies 64 bytes on each of
CUDA device 0 and CUDA device 1. Then 'usage' should have first two entries
set, one has the buffer attributes of "type GPU, id 0, 64 bytes" and the
other has "type GPU, id 1, 64 bytes". 'usage' is owned by the backend and
may be released after the function returns.usage_size
- The number of entries in 'usage'.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceReportMemoryUsage(TRITONBACKEND_ModelInstance instance, @ByPtrPtr TRITONSERVER_BufferAttributes usage, @Cast(value="uint32_t") int usage_size)
public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceReportStatistics(TRITONBACKEND_ModelInstance instance, TRITONBACKEND_Request request, @Cast(value="const bool") boolean success, @Cast(value="const uint64_t") long exec_start_ns, @Cast(value="const uint64_t") long compute_start_ns, @Cast(value="const uint64_t") long compute_end_ns, @Cast(value="const uint64_t") long exec_end_ns)
instance
- The model instance.request
- The inference request that statistics are being
reported for.success
- True if the inference request completed
successfully, false if it failed to complete.exec_start_ns
- Timestamp for the start of execution.compute_start_ns
- Timestamp for the start of execution
computations.compute_end_ns
- Timestamp for the end of execution
computations.exec_end_ns
- Timestamp for the end of execution.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceReportBatchStatistics(TRITONBACKEND_ModelInstance instance, @Cast(value="const uint64_t") long batch_size, @Cast(value="const uint64_t") long exec_start_ns, @Cast(value="const uint64_t") long compute_start_ns, @Cast(value="const uint64_t") long compute_end_ns, @Cast(value="const uint64_t") long exec_end_ns)
instance
- The model instance.batch_size
- Combined batch size of all the individual
requests executed in the batch.exec_start_ns
- Timestamp for the start of execution.compute_start_ns
- Timestamp for the start of execution
computations.compute_end_ns
- Timestamp for the end of execution
computations.exec_end_ns
- Timestamp for the end of execution.public static TRITONSERVER_Error TRITONBACKEND_Initialize(TRITONBACKEND_Backend backend)
Initialize a backend. This function is optional, a backend is not required to implement it. This function is called once when a backend is loaded to allow the backend to initialize any state associated with the backend. A backend has a single state that is shared across all models that use the backend.
backend
- The backend.public static TRITONSERVER_Error TRITONBACKEND_Finalize(TRITONBACKEND_Backend backend)
backend
- The backend.public static TRITONSERVER_Error TRITONBACKEND_ModelInitialize(TRITONBACKEND_Model model)
model
- The model.public static TRITONSERVER_Error TRITONBACKEND_ModelFinalize(TRITONBACKEND_Model model)
model
- The model.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceInitialize(TRITONBACKEND_ModelInstance instance)
instance
- The model instance.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceFinalize(TRITONBACKEND_ModelInstance instance)
instance
- The model instance.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceExecute(TRITONBACKEND_ModelInstance instance, @Cast(value="TRITONBACKEND_Request**") PointerPointer requests, @Cast(value="const uint32_t") int request_count)
instance
- The model instance.requests
- The requests.request_count
- The number of requests in the batch.public static TRITONSERVER_Error TRITONBACKEND_ModelInstanceExecute(TRITONBACKEND_ModelInstance instance, @ByPtrPtr TRITONBACKEND_Request requests, @Cast(value="const uint32_t") int request_count)
public static TRITONSERVER_Error TRITONBACKEND_GetBackendAttribute(TRITONBACKEND_Backend backend, TRITONBACKEND_BackendAttribute backend_attributes)
backend
- The backend.backend_attributes
- Return the backend attribute.public static TRITONSERVER_Error TRITONBACKEND_BackendAttributeAddPreferredInstanceGroup(TRITONBACKEND_BackendAttribute backend_attributes, @Cast(value="const TRITONSERVER_InstanceGroupKind") int kind, @Cast(value="const uint64_t") long count, @Cast(value="const uint64_t*") LongPointer device_ids, @Cast(value="const uint64_t") long id_count)
Add the preferred instance group of the backend. This function can be called multiple times to cover different instance group kinds that the backend supports, given the priority order that the first call describes the most preferred group. In the case where instance group are not explicitly provided, Triton will use this attribute to create model deployment that aligns more with the backend preference.
backend_attributes
- The backend attributes object.kind
- The kind of the instance group.count
- The number of instances per device. Triton default will be used
if 0 is provided.device_ids
- The devices where instances should be available. Triton
default will be used if 'nullptr' is provided.id_count
- The number of devices in 'device_ids'.public static TRITONSERVER_Error TRITONBACKEND_BackendAttributeAddPreferredInstanceGroup(TRITONBACKEND_BackendAttribute backend_attributes, @Cast(value="const TRITONSERVER_InstanceGroupKind") int kind, @Cast(value="const uint64_t") long count, @Cast(value="const uint64_t*") LongBuffer device_ids, @Cast(value="const uint64_t") long id_count)
public static TRITONSERVER_Error TRITONBACKEND_BackendAttributeAddPreferredInstanceGroup(TRITONBACKEND_BackendAttribute backend_attributes, @Cast(value="const TRITONSERVER_InstanceGroupKind") int kind, @Cast(value="const uint64_t") long count, @Cast(value="const uint64_t*") long[] device_ids, @Cast(value="const uint64_t") long id_count)
public static TRITONSERVER_Error TRITONBACKEND_BackendAttributeSetParallelModelInstanceLoading(TRITONBACKEND_BackendAttribute backend_attributes, @Cast(value="bool") boolean enabled)
backend_attributes
- The backend attributes object.enabled
- Whether or not the backend supports loading model instances
in parallel.public static TRITONSERVER_Error TRITONBACKEND_ModelBatcherInitialize(@Cast(value="TRITONBACKEND_Batcher**") PointerPointer batcher, TRITONBACKEND_Model model)
Create a new batcher for use with custom batching. This is called during model loading. The batcher will point to a user-defined data structure that holds read-only data used for custom batching.
batcher
- User-defined placeholder for backend to store and
retrieve information about the batching strategy for this
model.RITONBACKEND_ISPEC return a TRITONSERVER_Error indicating success or
failure. @param model The backend model for which Triton is forming a batch.public static TRITONSERVER_Error TRITONBACKEND_ModelBatcherInitialize(@ByPtrPtr TRITONBACKEND_Batcher batcher, TRITONBACKEND_Model model)
public static TRITONSERVER_Error TRITONBACKEND_ModelBatcherFinalize(TRITONBACKEND_Batcher batcher)
batcher
- User-defined placeholder for backend to store and
retrieve information about the batching strategy for this model.public static TRITONSERVER_Error TRITONBACKEND_ModelBatchIncludeRequest(TRITONBACKEND_Request request, Pointer userp, @Cast(value="bool*") boolean[] should_include)
request
- The request to be added to the pending batch.userp
- The placeholder for backend to store and retrieve information
about this pending batch. When the callback returns, this should reflect
the latest batch information.should_include
- The pointer to be updated on whether the request
should be included in the batch.public static TRITONSERVER_Error TRITONBACKEND_ModelBatchIncludeRequest(TRITONBACKEND_Request request, Pointer userp, @Cast(value="bool*") BoolPointer should_include)
public static TRITONSERVER_Error TRITONBACKEND_ModelBatchInitialize(@Const TRITONBACKEND_Batcher batcher, @Cast(value="void**") PointerPointer userp)
userp
- The placeholder for backend to store and retrieve information
/** about this pending batch.
/** @return a TRITONSERVER_Error indicating success or failure.public static TRITONSERVER_Error TRITONBACKEND_ModelBatchInitialize(@Const TRITONBACKEND_Batcher batcher, @Cast(value="void**") @ByPtrPtr Pointer userp)
public static TRITONSERVER_Error TRITONBACKEND_ModelBatchFinalize(Pointer userp)
userp
- The placeholder for backend to store and retrieve information
about this pending batch.public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, String name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") PointerPointer shape, @Cast(value="uint64_t*") LongPointer dim_count)
response
- The response object.name
- The name of the output.datatype
- Returns the type of the output.shape
- Returns the shape of the output.dim_count
- Returns the number of dimensions of the returned
shape.public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, String name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint64_t*") LongPointer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint64_t*") LongBuffer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, String name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint64_t*") long[] dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint64_t*") LongPointer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, String name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint64_t*") LongBuffer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutputByName(TRITONBACKEND_Response response, @Cast(value="const char*") BytePointer name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint64_t*") long[] dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutput(TRITONBACKEND_Response response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") PointerPointer shape, @Cast(value="uint64_t*") LongPointer dim_count)
response
- The response object.index
- The index of the output tensor, must be 0 <= index <
count, where 'count' is the value returned by
TRITONSERVER_InferenceResponseOutputCount.name
- Returns the name of the output.datatype
- Returns the type of the output.shape
- Returns the shape of the output.dim_count
- Returns the number of dimensions of the returned
shape.public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutput(TRITONBACKEND_Response response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer name, @Cast(value="TRITONSERVER_DataType*") IntPointer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongPointer shape, @Cast(value="uint64_t*") LongPointer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutput(TRITONBACKEND_Response response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer name, @Cast(value="TRITONSERVER_DataType*") IntBuffer datatype, @Cast(value="const int64_t**") @ByPtrPtr LongBuffer shape, @Cast(value="uint64_t*") LongBuffer dim_count)
public static TRITONSERVER_Error TRITONBACKEND_InferenceResponseOutput(TRITONBACKEND_Response response, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] name, @Cast(value="TRITONSERVER_DataType*") int[] datatype, @Cast(value="const int64_t**") @ByPtrPtr long[] shape, @Cast(value="uint64_t*") long[] dim_count)
public static TRITONSERVER_Error TRITONREPOAGENT_ApiVersion(@Cast(value="uint32_t*") IntPointer major, @Cast(value="uint32_t*") IntPointer minor)
major
- Returns the TRITONREPOAGENT API major version supported
by Triton.minor
- Returns the TRITONREPOAGENT API minor version supported
by Triton.public static TRITONSERVER_Error TRITONREPOAGENT_ApiVersion(@Cast(value="uint32_t*") IntBuffer major, @Cast(value="uint32_t*") IntBuffer minor)
public static TRITONSERVER_Error TRITONREPOAGENT_ApiVersion(@Cast(value="uint32_t*") int[] major, @Cast(value="uint32_t*") int[] minor)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocation(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="TRITONREPOAGENT_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") PointerPointer location)
agent
- The agent.model
- The model.artifact_type
- Returns the artifact type for the location.path
- Returns the location.public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocation(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="TRITONREPOAGENT_ArtifactType*") IntPointer artifact_type, @Cast(value="const char**") @ByPtrPtr BytePointer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocation(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="TRITONREPOAGENT_ArtifactType*") IntBuffer artifact_type, @Cast(value="const char**") @ByPtrPtr ByteBuffer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocation(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="TRITONREPOAGENT_ArtifactType*") int[] artifact_type, @Cast(value="const char**") @ByPtrPtr byte[] location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationAcquire(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, @Cast(value="const char**") PointerPointer location)
agent
- The agent.model
- The model.artifact_type
- The artifact type for the location.path
- Returns the location.public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationAcquire(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, @Cast(value="const char**") @ByPtrPtr BytePointer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationAcquire(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, @Cast(value="const char**") @ByPtrPtr ByteBuffer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationAcquire(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, @Cast(value="const char**") @ByPtrPtr byte[] location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationRelease(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, String location)
agent
- The agent.model
- The model.path
- The location to release.public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryLocationRelease(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const char*") BytePointer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryUpdate(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, String location)
agent
- The agent.model
- The model.artifact_type
- The artifact type for the location.path
- Returns the location.public static TRITONSERVER_Error TRITONREPOAGENT_ModelRepositoryUpdate(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ArtifactType") int artifact_type, @Cast(value="const char*") BytePointer location)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameterCount(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="uint32_t*") IntPointer count)
agent
- The agent.model
- The model.count
- Returns the number of input tensors.public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameterCount(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="uint32_t*") IntBuffer count)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameterCount(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="uint32_t*") int[] count)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameter(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int index, @Cast(value="const char**") PointerPointer parameter_name, @Cast(value="const char**") PointerPointer parameter_value)
agent
- The agent.model
- The model.index
- The index of the parameter. Must be 0 <= index <
count, where count is the value returned by
TRITONREPOAGENT_ModelParameterCount.parameter_name
- Returns the name of the parameter.parameter_value
- Returns the value of the parameter.public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameter(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr BytePointer parameter_name, @Cast(value="const char**") @ByPtrPtr BytePointer parameter_value)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameter(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr ByteBuffer parameter_name, @Cast(value="const char**") @ByPtrPtr ByteBuffer parameter_value)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelParameter(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int index, @Cast(value="const char**") @ByPtrPtr byte[] parameter_name, @Cast(value="const char**") @ByPtrPtr byte[] parameter_value)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelConfig(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int config_version, @Cast(value="TRITONSERVER_Message**") PointerPointer model_config)
agent
- The agent.model
- The model.config_version
- The model configuration will be returned in
a format matching this version. If the configuration cannot be
represented in the requested version's format then an error will
be returned. Currently only version 1 is supported.model_config
- Returns the model configuration as a message.public static TRITONSERVER_Error TRITONREPOAGENT_ModelConfig(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const uint32_t") int config_version, @ByPtrPtr TRITONSERVER_Message model_config)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelState(TRITONREPOAGENT_AgentModel model, @Cast(value="void**") PointerPointer state)
model
- The agent model.state
- Returns the user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONREPOAGENT_ModelState(TRITONREPOAGENT_AgentModel model, @Cast(value="void**") @ByPtrPtr Pointer state)
public static TRITONSERVER_Error TRITONREPOAGENT_ModelSetState(TRITONREPOAGENT_AgentModel model, Pointer state)
model
- The agent model.state
- The user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONREPOAGENT_State(TRITONREPOAGENT_Agent agent, @Cast(value="void**") PointerPointer state)
agent
- The agent.state
- Returns the user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONREPOAGENT_State(TRITONREPOAGENT_Agent agent, @Cast(value="void**") @ByPtrPtr Pointer state)
public static TRITONSERVER_Error TRITONREPOAGENT_SetState(TRITONREPOAGENT_Agent agent, Pointer state)
agent
- The agent.state
- The user state, or nullptr if no user state.public static TRITONSERVER_Error TRITONREPOAGENT_Initialize(TRITONREPOAGENT_Agent agent)
Initialize an agent. This function is optional. This function is called once when an agent is loaded to allow the agent to initialize any state associated with the agent. An agent has a single state that is shared across all invocations of the agent.
agent
- The agent.public static TRITONSERVER_Error TRITONREPOAGENT_Finalize(TRITONREPOAGENT_Agent agent)
agent
- The agent.public static TRITONSERVER_Error TRITONREPOAGENT_ModelInitialize(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model)
agent
- The agent to be associated with the model.model
- The model.public static TRITONSERVER_Error TRITONREPOAGENT_ModelFinalize(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model)
agent
- The agent associated with the model.model
- The model.public static TRITONSERVER_Error TRITONREPOAGENT_ModelAction(TRITONREPOAGENT_Agent agent, TRITONREPOAGENT_AgentModel model, @Cast(value="const TRITONREPOAGENT_ActionType") int action_type)
agent
- The agent.model
- The model that is the target of the action.
\action_type The type of action the agent should handle for the model.Copyright © 2024. All rights reserved.