This is a dedicated watch page for a single video.
A company needs to process a large inference payload of several gigabytes (GBs) for its machine learning model built using Amazon SageMaker. The company's systems can tolerate some waiting time for the responses. Which inference method would be the most suitable for this scenario?