scalable batch inference for millions of time series (SKU sales, 1000-day horizon) on SageMaker #368
skwskwskwskw
started this conversation in
Ideas
Replies: 1 comment
-
|
@skwskwskwskw I think any of the options that you propose should be good enough because the Chronos-2 model is extremely fast. My recommendation would be to deploy the model on Sagemaker Jumpstart as shown in this notebook. This would likely be enough. You could also use the jumpstart model for a batch-transform job. For comparisons to other models, you can check fev-bench and GIFT-Eval. I also think that a long forecast horizon of 1000 days is likely not a great idea. You would be better off aggregating the data and forecasting at a coarser granularity (weekly or monthly) instead of using such a long prediction horizon. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Request:
Context
I need to run forecast inference for millions of SKU-level time series (SKU × location/product hierarchy). The forecast horizon is 1,000 days. I’m looking for guidance on a cost-efficient and operationally sound approach on Amazon SageMaker.
Questions
Throughput & cost efficiency
Naïve per-series loops won’t finish in a reasonable time. What’s the most cost-effective pattern on SageMaker for large-scale batch inference here (e.g., Batch Transform vs. distributed inference on Processing jobs, SageMaker Inference Pipelines, Async endpoints, or other recommended patterns)?
High cardinality (SKU × hierarchy × location)
The number of series explodes due to product hierarchy and geography. Any best practices to partition/shard the workload and optimize parallelism (data layout in S3, instance types/count, I/O considerations) to keep runtime and cost under control?
Method comparison / benchmarks
My current solution uses Nixtla’s
mlforecastwith LightGBM (mlforecast+lightgbm). Are there benchmarks or case studies comparing this setup with your recommended approach (accuracy, latency, and cost) at a similar scale?Environment & constraints (if helpful)
mlforecast+lightgbmWhat I’m hoping for
mlforecast-LightGBM to your method at high cardinalityThanks in advance for any pointers, designs, or examples!
Beta Was this translation helpful? Give feedback.
All reactions