|
| 1 | +"""Callback Handler that tracks AIMessage.usage_metadata.""" |
| 2 | + |
| 3 | +import threading |
| 4 | +from collections.abc import Generator |
| 5 | +from contextlib import contextmanager |
| 6 | +from contextvars import ContextVar |
| 7 | +from typing import Any, Optional |
| 8 | + |
| 9 | +from langchain_core.callbacks import BaseCallbackHandler |
| 10 | +from langchain_core.messages import AIMessage |
| 11 | +from langchain_core.messages.ai import UsageMetadata, add_usage |
| 12 | +from langchain_core.outputs import ChatGeneration, LLMResult |
| 13 | + |
| 14 | + |
| 15 | +class UsageMetadataCallbackHandler(BaseCallbackHandler): |
| 16 | + """Callback Handler that tracks AIMessage.usage_metadata. |
| 17 | +
|
| 18 | + Example: |
| 19 | + .. code-block:: python |
| 20 | +
|
| 21 | + from langchain.chat_models import init_chat_model |
| 22 | + from langchain_core.callbacks import UsageMetadataCallbackHandler |
| 23 | +
|
| 24 | + llm = init_chat_model(model="openai:gpt-4o-mini") |
| 25 | +
|
| 26 | + callback = UsageMetadataCallbackHandler() |
| 27 | + results = llm.batch(["Hello", "Goodbye"], config={"callbacks": [callback]}) |
| 28 | + print(callback.usage_metadata) |
| 29 | +
|
| 30 | + .. code-block:: none |
| 31 | +
|
| 32 | + {'output_token_details': {'audio': 0, 'reasoning': 0}, 'input_tokens': 17, 'output_tokens': 31, 'total_tokens': 48, 'input_token_details': {'cache_read': 0, 'audio': 0}} |
| 33 | +
|
| 34 | + .. versionadded:: 0.3.49 |
| 35 | + """ # noqa: E501 |
| 36 | + |
| 37 | + def __init__(self) -> None: |
| 38 | + super().__init__() |
| 39 | + self._lock = threading.Lock() |
| 40 | + self.usage_metadata: Optional[UsageMetadata] = None |
| 41 | + |
| 42 | + def __repr__(self) -> str: |
| 43 | + return str(self.usage_metadata) |
| 44 | + |
| 45 | + def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: |
| 46 | + """Collect token usage.""" |
| 47 | + # Check for usage_metadata (langchain-core >= 0.2.2) |
| 48 | + try: |
| 49 | + generation = response.generations[0][0] |
| 50 | + except IndexError: |
| 51 | + generation = None |
| 52 | + if isinstance(generation, ChatGeneration): |
| 53 | + try: |
| 54 | + message = generation.message |
| 55 | + if isinstance(message, AIMessage): |
| 56 | + usage_metadata = message.usage_metadata |
| 57 | + else: |
| 58 | + usage_metadata = None |
| 59 | + except AttributeError: |
| 60 | + usage_metadata = None |
| 61 | + else: |
| 62 | + usage_metadata = None |
| 63 | + |
| 64 | + # update shared state behind lock |
| 65 | + with self._lock: |
| 66 | + self.usage_metadata = add_usage(self.usage_metadata, usage_metadata) |
| 67 | + |
| 68 | + |
| 69 | +@contextmanager |
| 70 | +def get_usage_metadata_callback( |
| 71 | + name: str = "usage_metadata_callback", |
| 72 | +) -> Generator[UsageMetadataCallbackHandler, None, None]: |
| 73 | + """Get context manager for tracking usage metadata across chat model calls using |
| 74 | + ``AIMessage.usage_metadata``. |
| 75 | +
|
| 76 | + Args: |
| 77 | + name (str): The name of the context variable. Defaults to |
| 78 | + ``"usage_metadata_callback"``. |
| 79 | +
|
| 80 | + Example: |
| 81 | + .. code-block:: python |
| 82 | +
|
| 83 | + from langchain.chat_models import init_chat_model |
| 84 | + from langchain_core.callbacks import get_usage_metadata_callback |
| 85 | +
|
| 86 | + llm = init_chat_model(model="openai:gpt-4o-mini") |
| 87 | +
|
| 88 | + with get_usage_metadata_callback() as cb: |
| 89 | + llm.invoke("Hello") |
| 90 | + llm.invoke("Goodbye") |
| 91 | + print(cb.usage_metadata) |
| 92 | +
|
| 93 | + .. code-block:: none |
| 94 | +
|
| 95 | + {'output_token_details': {'audio': 0, 'reasoning': 0}, 'input_tokens': 17, 'output_tokens': 31, 'total_tokens': 48, 'input_token_details': {'cache_read': 0, 'audio': 0}} |
| 96 | +
|
| 97 | + .. versionadded:: 0.3.49 |
| 98 | + """ # noqa: E501 |
| 99 | + from langchain_core.tracers.context import register_configure_hook |
| 100 | + |
| 101 | + usage_metadata_callback_var: ContextVar[Optional[UsageMetadataCallbackHandler]] = ( |
| 102 | + ContextVar(name, default=None) |
| 103 | + ) |
| 104 | + register_configure_hook(usage_metadata_callback_var, True) |
| 105 | + cb = UsageMetadataCallbackHandler() |
| 106 | + usage_metadata_callback_var.set(cb) |
| 107 | + yield cb |
| 108 | + usage_metadata_callback_var.set(None) |
0 commit comments