-
Notifications
You must be signed in to change notification settings - Fork 736
docs(amazonq): Add latency metric diagrams for amazon q chat #5865
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
docs/telemetry-perf.md
Outdated
|
|
||
| ### cwsprChatTimeToFirstChunk | ||
|
|
||
| The time between when the conversation id was created and when we got back the first usable result |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
when the conversation id was created
Can this be reworded to something higher level like "when the first conversation message is sent" (IIUC)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Changed to "when the conversation stream was created"
| participant user as User | ||
| participant chat as Chat UI | ||
| participant vscode as VSCode | ||
| participant generateResponse as Generate response |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is this the service client?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Its technically the name of the function that executes the calls to the backend, I wanted it to live seperate from the vscode extension host since the chat message -> vscode extension host happens completely seperately from that function. Let me know if theres a better name you suggest
| end | ||
| ``` | ||
|
|
||
| ### cwsprChatTimeBetweenChunks |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we clarify somewhere what a "chunk" is in a chat context?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I've changed the description to include
"successive pieces" of data rather than "successive chunks". Does that make more sense?
Problem
We have no documentation showing what latencies we are measuring for amazon q chat
Solution
Add documentation
Direct link: https://github.com/jpinkney-aws/aws-toolkit-vscode/blob/4c7ae0ae0c8d0e4525b68da487bd43c4c1b7884b/docs/telemetry-perf.md
License: I confirm that my contribution is made under the terms of the Apache 2.0 license.