This is a minimalist, performance-focused tracing framework for building Python applications, particularly those involving LLM chains. We built this while refining LLM applications to be production-ready and didn't want to wait for other solutions out there. Feel free to file issues or provide feedback: sudowoodo200 [at] gmail [dot] com.
- Install from PyPI:
pip install impulse-core - Direct install:
make install. Note that this is not an editable installation
This initializes the objects with default settings
tracerwill use aLocalLogger, which writes json records to a file at.impulselogs/logs_{timestamp}.json- Currently, this also supports logging to a MongoDB database out of the box. Use
MongoLoggerclass instead. (See tutorial) tracer.hook()will be set to the default thread at"default"
from impulse_core import ImpulseTracer
tracer = ImpulseTracer()
@tracer.hook()
def some_function(x: int, y: str = 2) -> str:
return f"{str(x)} - {y}"
def handle_request():
tracer.set_session_id("user_abc_session_1")
some_function(1)
tracer.shutdown() ## needed for local logger to flush the write bufferThe record will capture information (under the "payload" field of the json record) during the function call:
{
"function": {
"name" : "some_function",
...
},
"trace_module": {
"tracer_id": "asdfase-234234sdafs-aerwer",
"thread_id": "default",
"hook_id": "some_function",
...
}
"call_id": "asfda2323-52sdfasd",
"timestamps": {
"start": "2023-08-20 22:05:55.000000",
"end": "2023-08-20 22:05:56.123456",
"start_to_end_seconds": "1.123456"
},
"arguments": {
"x": 1,
"y": 2
},
"status": "success",
"output": "1 - 2",
...
}Each record is uniquely identified by 4 fields:
- A
callis every single run of thetraced function, identified by acall_idfield in the logs. Eachcallalso defines atrace_log()context. (see below) - A
hookis a decorator for a specific function, identified byhook_idargument in the@tracer.hook()function. - A
threadis a collection ofhook's, identified by thethread_idargument in the@tracer.hook()function. - A
moduleis an instance of theImpulseTracerclass, identified by theinstance_idattribute and manages a collection ofthreads
This works with functions, methods, classmethods, staticmethods, coroutines, and async generators. If an exception occurs, logging will still happen.
You can trace nested calls by decorating the relevant functions. For instance:
@tracer.hook()
def top_level():
return intermediate()
def intermediate():
return some_function(1,1)The log records will preserve the parent-child relationship between some_function(x,y) and top_level() in the stack_trace field. For instance, this will be captured in the top_level()'s record:
{
"function": {
"name" : "top_level",
...
},
"call_id": "asdfasdf-2352dsafsa",
...
"stack_trace": {
"parents": [
...
],
"children": [
{
"fn_name": "some_function",
"call_id": "asfda2323-52sdfasd" ,
"trace_module": { ... }
},
]
}
...
}
{
"function": {
"name" : "some_function",
...
},
"call_id": "asfda2323-52sdfasd",
...
"stack_trace": {
"parents": [
{
"fn_name": "top_level",
"call_id": "asdfasdf-2352dsafsa",
"trace_module": { ... }
},
],
"children": [
...
]
}
// ...
}Each @trace.hook() creates a context until superceded by a nested hook.
Another simple by powerful feature is the ability to log arbitrary data, timestamped, directly into the context, which is then included as part of the enclosing logging record. The only restriction is that it must be convertible with json.dumps.
from impulse_core import trace_log as log
@tracer.hook()
def some_function(x: int, y: str = 2) -> str:
log("The ents shall march to")
log({"location": "Isengard"})
return f"{str(x)} - {y}"These can be accessed in the "trace_log" field of the record.
{
"function": {
"name" : "some_function",
...
},
...
"trace_log": [
{
"timestamp": "2023-08-20 22:05:55.000511",
"payload": "The ents shall march to"
},
...
]
}Common use cases include capturing session data when serving web requests and doing more granular logging of function components.
Apologies for the lack of docs for now. Still drafting it. In its place, a quick tutorial can be found at app/tutorial/tutorial.ipynb. To get started, use the following to boot up a local instance of a database and a (very rough) exploration app in Streamlit
make appYou can also only boot up the database without the Streamlit app.
make databaseAfter you are done, clean up the assets with
make shutdownDetailed overview of the logging schema can be found at impulse_core/schema.py