Prefix caching is a key factor in LLM inference performance, but there's currently no visibility into cache behavior in the aiperf output.