-
Notifications
You must be signed in to change notification settings - Fork 182
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Do we need to distinguish client side and server side llm call? #1079
Comments
Adding @SergeyKanzhelev who might be interested in GCP server LLM metrics. The assumption is that client and server would have different information available. E.g. Client and server metrics might have different attributes. E.g. server might have information about pricing tier, region, availability zone that is not available on the client, but are very useful to know. Therefore we either need:
Other otel semantic conventions use option 1 and this is a good reason for |
@lmolkova @nirga there is an issue tracking the vLLM metrics vllm-project/vllm#5041 and also a metrics proposal at https://docs.google.com/document/d/1SpSp1E6moa4HSrJnS4x3NpLuj88sMXr2tbofKlzTZpk/edit?resourcekey=0-ob5dR-AJxLQ5SvPlA4rdsg#heading=h.qmzyorj64um1 |
I believe this can be resolved:
@gyliu513 please comment if you believe there is something else we need to do on this issue (and feel free to reopen it) |
Area(s)
area:gen-ai
What happened?
Description
There is a PR trying to enable vllm support metrics as well and it is adopting this semantic convention as well vllm-project/vllm#4687 , here the question is vllm is server side, but I can see seems now the llm semantic convention are manily trying to instrument client side code like https://github.com/traceloop/openllmetry/tree/main/packages, do we need to distinguish the client side and server side semantic convetion? Thanks!
@nirga @lmolkova ^^
Semantic convention version
NA
Additional context
No response
The text was updated successfully, but these errors were encountered: