Source code for oumi.core.configs.inference_config
# Copyright 2025 - Oumi## Licensed under the Apache License, Version 2.0 (the "License");# you may not use this file except in compliance with the License.# You may obtain a copy of the License at## http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing, software# distributed under the License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.# See the License for the specific language governing permissions and# limitations under the License.fromdataclassesimportdataclass,fieldfromtypingimportOptionalfromoumi.core.configs.base_configimportBaseConfigfromoumi.core.configs.inference_engine_typeimportInferenceEngineTypefromoumi.core.configs.params.generation_paramsimportGenerationParamsfromoumi.core.configs.params.model_paramsimportModelParamsfromoumi.core.configs.params.remote_paramsimportRemoteParams
[docs]@dataclassclassInferenceConfig(BaseConfig):model:ModelParams=field(default_factory=ModelParams)"""Parameters for the model used in inference."""generation:GenerationParams=field(default_factory=GenerationParams)"""Parameters for text generation during inference."""input_path:Optional[str]=None"""Path to the input file containing prompts for text generation. The input file should be in JSONL format, where each line is a JSON representation of an Oumi `Conversation` object. """output_path:Optional[str]=None"""Path to the output file where the generated text will be saved."""engine:Optional[InferenceEngineType]=None"""The inference engine to use for generation. Options: - NATIVE: Use the native inference engine via a local forward pass. - VLLM: Use the vLLM inference engine started locally by oumi. - REMOTE_VLLM: Use the external vLLM inference engine. - SGLANG: Use the SGLang inference engine. - LLAMACPP: Use LlamaCPP inference engine. - REMOTE: Use the inference engine for APIs that implement the OpenAI Chat API interface. - ANTHROPIC: Use the inference engine for Anthropic's API. If not specified, the "NATIVE" engine will be used. """remote_params:Optional[RemoteParams]=None"""Parameters for running inference against a remote API."""