Replies: 10 comments 12 replies
-
|
There is no direct information in the available sources about troubleshooting slow UI latency for viewing traces or single trace details in a self-hosted Langfuse setup with your described AWS/ClickHouse configuration. The documentation covers how traces are queried and displayed, and describes that the Langfuse UI and APIs rely on aggregating and querying trace data, often involving complex queries, especially for large datasets(1)(2)(3). There is also mention that the query engine is designed for flexibility and performance(2). However, there are no explicit guidelines or troubleshooting steps for performance tuning or diagnosing latency issues in your specific architecture. If you would like, I can try to provide a general answer based on typical Langfuse and ClickHouse performance practices. Check out these potentially useful sources: Would you like a general answer based on best practices for performance in Langfuse deployments? 📚 Sources: Have another question? Just tag @inkeep. |
Beta Was this translation helpful? Give feedback.
-
|
So i wiped out my clickhouse and reset back to stock and the performance is back to normal. I had emitted about 10k traces before this. |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
Asked ChatGPT |
Beta Was this translation helpful? Give feedback.
-
|
Hi @dahnny012, we have docs on sizing and sclaing Langfuse here: https://langfuse.com/self-hosting/configuration/scaling |
Beta Was this translation helpful? Give feedback.
-
|
Reaching out to @maxdeichmann because of https://langfuse.com/changelog/2024-02-29-performance-improvement-ui |
Beta Was this translation helpful? Give feedback.
-
|
Running in a sagemaker notebook with the same specs but using EBS storage results in a 100x speed up. Need to see how this scales over time. |
Beta Was this translation helpful? Give feedback.
-
|
After transitioning the infra to use ec2 with EBS it is so much faster. |
Beta Was this translation helpful? Give feedback.
-
|
Latency is in the 100-300ms range if you adjust it from EFS bursting to EFS Elastic. |
Beta Was this translation helpful? Give feedback.
-
|
I also solved this by moving from EFS Bursting to EFS Enhanced -> Elastic |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Config:
--max-old-space-size=${memoryLimitMiB}, on my langfuse ecs container, set ALB timeout to +10sProblem:
I get abysmal latency on the UI when getting the list of traces in the last 24 hours. Some of the rows wont populate and loading a single trace takes on average 15s. In AWS I'm not seeing any metrics higher than 17% for RAM and CPU usage.
This issue started happening as we put in more traces into the system. We're currently at 6gb of traces (based on EFS)
Beta Was this translation helpful? Give feedback.
All reactions