Source code for ray.serve.llm.configs
from ray.llm._internal.serve.configs.server_models import (
LLMConfig as _LLMConfig,
LLMServingArgs as _LLMServingArgs,
ModelLoadingConfig as _ModelLoadingConfig,
CloudMirrorConfig as _CloudMirrorConfig,
LoraConfig as _LoraConfig,
)
from ray.util.annotations import PublicAPI
[docs]
@PublicAPI(stability="alpha")
class LLMConfig(_LLMConfig):
"""The configuration for starting an LLM deployment."""
pass
[docs]
@PublicAPI(stability="alpha")
class LLMServingArgs(_LLMServingArgs):
"""The configuration for starting an LLM deployment application."""
pass
[docs]
@PublicAPI(stability="alpha")
class ModelLoadingConfig(_ModelLoadingConfig):
"""The configuration for loading an LLM model."""
pass
[docs]
@PublicAPI(stability="alpha")
class CloudMirrorConfig(_CloudMirrorConfig):
"""The configuration for mirroring an LLM model from cloud storage."""
pass
[docs]
@PublicAPI(stability="alpha")
class LoraConfig(_LoraConfig):
"""The configuration for loading an LLM model with LoRA."""
pass