Interface InferenceEngine<T>

    • Method Detail

      • initialize

        void initialize​(ModelConfig config)
                 throws InferenceException
        Initializes the inference engine with the given configuration.

        Loads the model and prepares the engine for inference operations.

        Parameters:
        config - model configuration
        Throws:
        InferenceException - if initialization fails
      • inferBatch

        InferenceResult inferBatch​(Map<String,​Object>[] batchInputs)
                            throws InferenceException
        Performs batch inference on multiple inputs.

        Batch inference is typically more efficient than multiple single inferences.

        Parameters:
        batchInputs - array of input maps
        Returns:
        inference result containing batch predictions
        Throws:
        InferenceException - if inference fails
      • getCapabilities

        InferenceEngine.EngineCapabilities getCapabilities()
        Gets the capabilities of this inference engine.
        Returns:
        engine capabilities (batching, GPU support, etc.)
      • close

        void close()
            throws InferenceException
        Closes the inference engine and releases all resources.

        After calling this method, the engine should not be used again.

        Throws:
        InferenceException - if cleanup fails
      • isReady

        boolean isReady()
        Checks if the engine is ready for inference operations.
        Returns:
        true if engine is initialized and ready
      • getMetadata

        ModelMetadata getMetadata()
        Gets metadata about the loaded model.
        Returns:
        model metadata including inputs, outputs, and format
      • getModelConfig

        ModelConfig getModelConfig()
        Gets the configuration used to initialize this engine.
        Returns:
        model configuration