A media company wants to process large volumes of real-time data from its streaming platform and store the processed data in an S3-based data lake. The company needs a cost-effective and scalable ETL solution that can run Spark jobs and automatically scale based on the amount of data being processed. The team also requires the ability to adjust the compute resources for the ETL jobs to optimize costs. Which AWS service and configuration should they choose?