Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion llama_stack/cli/stack/run.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@

REPO_ROOT = Path(__file__).parent.parent.parent.parent

logger = get_logger(name=__name__, category="server")
logger = get_logger(name=__name__, category="cli")


class StackRun(Subcommand):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/datasets.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.datatypes import RoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routers")


class DatasetIORouter(DatasetIO):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/eval_scoring.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.datatypes import RoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routers")


class ScoringRouter(Scoring):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/inference.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@
from llama_stack.providers.utils.inference.inference_store import InferenceStore
from llama_stack.providers.utils.telemetry.tracing import get_current_span

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="core::routers")


class InferenceRouter(Inference):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/safety.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.datatypes import RoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routers")


class SafetyRouter(Safety):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/tool_runtime.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@

from ..routing_tables.toolgroups import ToolGroupsRoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routers")


class ToolRuntimeRouter(ToolRuntime):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routers/vector_io.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.datatypes import HealthResponse, HealthStatus, RoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routers")


class VectorIORouter(VectorIO):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/benchmarks.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@

from .common import CommonRoutingTableImpl

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class BenchmarksRoutingTable(CommonRoutingTableImpl, Benchmarks):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/common.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.datatypes import Api, RoutingTable

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


def get_impl_api(p: Any) -> Api:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/datasets.py
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@

from .common import CommonRoutingTableImpl

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class DatasetsRoutingTable(CommonRoutingTableImpl, Datasets):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/models.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@

from .common import CommonRoutingTableImpl, lookup_model

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class ModelsRoutingTable(CommonRoutingTableImpl, Models):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/scoring_functions.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@

from .common import CommonRoutingTableImpl

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class ScoringFunctionsRoutingTable(CommonRoutingTableImpl, ScoringFunctions):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/shields.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@

from .common import CommonRoutingTableImpl

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class ShieldsRoutingTable(CommonRoutingTableImpl, Shields):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/toolgroups.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@

from .common import CommonRoutingTableImpl

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


def parse_toolgroup_from_toolgroup_name_pair(toolgroup_name_with_maybe_tool_name: str) -> str | None:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/routing_tables/vector_dbs.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@

from .common import CommonRoutingTableImpl, lookup_model

logger = get_logger(name=__name__, category="core")
logger = get_logger(name=__name__, category="core::routing_tables")


class VectorDBsRoutingTable(CommonRoutingTableImpl, VectorDBs):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/server/auth.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
from llama_stack.core.server.routes import find_matching_route, initialize_route_impls
from llama_stack.log import get_logger

logger = get_logger(name=__name__, category="auth")
logger = get_logger(name=__name__, category="core::auth")


class AuthenticationMiddleware:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/server/auth_providers.py
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
)
from llama_stack.log import get_logger

logger = get_logger(name=__name__, category="auth")
logger = get_logger(name=__name__, category="core::auth")


class AuthResponse(BaseModel):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/server/quota.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
from llama_stack.providers.utils.kvstore.config import KVStoreConfig, SqliteKVStoreConfig
from llama_stack.providers.utils.kvstore.kvstore import kvstore_impl

logger = get_logger(name=__name__, category="quota")
logger = get_logger(name=__name__, category="core::server")


class QuotaMiddleware:
Expand Down
4 changes: 2 additions & 2 deletions llama_stack/core/server/server.py
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@

REPO_ROOT = Path(__file__).parent.parent.parent.parent

logger = get_logger(name=__name__, category="server")
logger = get_logger(name=__name__, category="core::server")


def warn_with_traceback(message, category, filename, lineno, file=None, line=None):
Expand Down Expand Up @@ -415,7 +415,7 @@ def main(args: argparse.Namespace | None = None):
config_contents = yaml.safe_load(fp)
if isinstance(config_contents, dict) and (cfg := config_contents.get("logging_config")):
logger_config = LoggingConfig(**cfg)
logger = get_logger(name=__name__, category="server", config=logger_config)
logger = get_logger(name=__name__, category="core::server", config=logger_config)
if args.env:
for env_pair in args.env:
try:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/store/registry.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@
from llama_stack.providers.utils.kvstore import KVStore, kvstore_impl
from llama_stack.providers.utils.kvstore.config import KVStoreConfig, SqliteKVStoreConfig

logger = get_logger(__name__, category="core")
logger = get_logger(__name__, category="core::registry")


class DistributionRegistry(Protocol):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/core/utils/config_resolution.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@
from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR
from llama_stack.log import get_logger

logger = get_logger(name=__name__, category="config_resolution")
logger = get_logger(name=__name__, category="core")


DISTRO_DIR = Path(__file__).parent.parent.parent.parent / "llama_stack" / "distributions"
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/models/llama/llama3/multimodal/model.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@

MP_SCALE = 8

logger = get_logger(name=__name__, category="models")
logger = get_logger(name=__name__, category="models::llama")


def reduce_from_tensor_model_parallel_region(input_):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/models/llama/llama3/tool_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@

from ..datatypes import BuiltinTool, RecursiveType, ToolCall, ToolPromptFormat

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="models::llama")

BUILTIN_TOOL_PATTERN = r'\b(?P<tool_name>\w+)\.call\(query="(?P<query>[^"]*)"\)'
CUSTOM_TOOL_CALL_PATTERN = re.compile(r"<function=(?P<function_name>[^}]+)>(?P<args>{.*?})")
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/models/llama/llama4/quantization/loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
from ..model import Transformer, TransformerBlock
from ..moe import MoE

log = get_logger(name=__name__, category="models")
log = get_logger(name=__name__, category="models::llama")


def swiglu_wrapper_no_reduce(
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/models/llama/quantize_impls.py
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@

from llama_stack.log import get_logger

log = get_logger(name=__name__, category="llama")
log = get_logger(name=__name__, category="models::llama")

try:
import fbgemm_gpu.experimental.gen_ai # noqa: F401
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@ def make_random_string(length: int = 8):
WEB_SEARCH_TOOL = "web_search"
RAG_TOOL_GROUP = "builtin::rag"

logger = get_logger(name=__name__, category="agents")
logger = get_logger(name=__name__, category="agents::meta_reference")


class ChatAgent(ShieldRunnerMixin):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@
from .persistence import AgentInfo
from .responses.openai_responses import OpenAIResponsesImpl

logger = get_logger(name=__name__, category="agents")
logger = get_logger(name=__name__, category="agents::meta_reference")


class MetaReferenceAgentsImpl(Agents):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.utils.kvstore import KVStore

log = get_logger(name=__name__, category="agents")
log = get_logger(name=__name__, category="agents::meta_reference")


class AgentSessionInfo(Session):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@
convert_response_text_to_chat_response_format,
)

logger = get_logger(name=__name__, category="responses")
logger = get_logger(name=__name__, category="openai::responses")


class OpenAIResponsePreviousResponseWithInputItems(BaseModel):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@
from .types import ChatCompletionContext, ChatCompletionResult
from .utils import convert_chat_choice_to_response_message, is_function_tool_call

logger = get_logger(name=__name__, category="responses")
logger = get_logger(name=__name__, category="agents::meta_reference")


class StreamingResponseOrchestrator:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@

from .types import ChatCompletionContext, ToolExecutionResult

logger = get_logger(name=__name__, category="responses")
logger = get_logger(name=__name__, category="agents::meta_reference")


class ToolExecutor:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
from llama_stack.log import get_logger
from llama_stack.providers.utils.telemetry import tracing

log = get_logger(name=__name__, category="agents")
log = get_logger(name=__name__, category="agents::meta_reference")


class SafetyException(Exception): # noqa: N818
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@
from .config import FireworksImplConfig
from .models import MODEL_ENTRIES

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::fireworks")


class FireworksInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProviderData):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@

from .models import MODEL_ENTRIES

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::llama_openai_compat")


class LlamaCompatInferenceAdapter(OpenAIMixin, LiteLLMOpenAIMixin):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/nvidia/nvidia.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,7 @@
)
from .utils import _is_nvidia_hosted

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::nvidia")


class NVIDIAInferenceAdapter(OpenAIMixin, Inference, ModelRegistryHelper):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/nvidia/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@

from . import NVIDIAConfig

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::nvidia")


def _is_nvidia_hosted(config: NVIDIAConfig) -> bool:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/ollama/ollama.py
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@

from .models import MODEL_ENTRIES

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::ollama")


class OllamaInferenceAdapter(
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/openai/openai.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
from .config import OpenAIConfig
from .models import MODEL_ENTRIES

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::openai")


#
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/tgi/tgi.py
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@

from .config import InferenceAPIImplConfig, InferenceEndpointImplConfig, TGIImplConfig

log = get_logger(name=__name__, category="inference")
log = get_logger(name=__name__, category="inference::tgi")


def build_hf_repo_model_entries():
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@
from .config import TogetherImplConfig
from .models import MODEL_ENTRIES

logger = get_logger(name=__name__, category="inference")
logger = get_logger(name=__name__, category="inference::together")


class TogetherInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProviderData):
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/inference/vllm/vllm.py
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@

from .config import VLLMInferenceAdapterConfig

log = get_logger(name=__name__, category="inference")
log = get_logger(name=__name__, category="inference::vllm")


def build_hf_repo_model_entries():
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@

from .config import NvidiaPostTrainingConfig

logger = get_logger(name=__name__, category="integration")
logger = get_logger(name=__name__, category="post_training::nvidia")


def warn_unsupported_params(config_dict: Any, supported_keys: set[str], config_name: str) -> None:
Expand Down
2 changes: 1 addition & 1 deletion llama_stack/providers/remote/safety/bedrock/bedrock.py
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@

from .config import BedrockSafetyConfig

logger = get_logger(name=__name__, category="safety")
logger = get_logger(name=__name__, category="safety::bedrock")


class BedrockSafetyAdapter(Safety, ShieldsProtocolPrivate):
Expand Down
Loading
Loading