High latency observed calling feature store and SageMaker service from EKS deployed Java application #4137
tushar-meesho
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We have recently done an integration with AWS SageMaker and FeatureStore for one of our use-case. On Cloudwatch, we could see the p99: 15-20ms for sagemaker and p99: 30-40ms for feature-store.
But we are observing the high latencies while using JAVA AWS sdk 2.0, to call to feature-store, somewhat around. p99: 150ms for sagemaker and p99: 100ms for feature-store.
We have the containerised spring-boot application running on EKS and here we are calling FS and SM.
Sample code :-
feature-store call
Sagemaker call:
Actual
SageMaker
call method and instrumentation over it.Using AWS java-sdk2 for SageMaker and feature-store.
On cloudwatch for Sagemaker we are using: p99 of ModelLatency + OverheadLatency metric as stated in the docs : docs.aws.amazon.com/sagemaker/latest/dg/…
.And For feature-store we are using Latency metrics as mentioned.
We are benchmarking our application latency from these metrics, but there's a significant difference in the latencies observed on the application side.
This is how we are measuring the end to end p99 latency on application level for sagemaker and feature-store:
histogram_quantile(0.99, sum(rate(resilience4j_circuitbreaker_calls_seconds_bucket{name="awsSageMaker"}[1m])) by (le))
. and same query with method name awsFeatureMaker as well.Could this be because of different VPC(EKS pods and SageMaker both are in the same region though AP_SOUTHEAST_1), or some other form of network latency, etc ? Is there a way to figure the cause of this ?
Beta Was this translation helpful? Give feedback.
All reactions