Become Amazon Certified with updated Data-Engineer-Associate exam questions and correct answers
Files from multiple data sources arrive in an Amazon S3 bucket on a regular basis. A data engineer wants to ingest new files into Amazon Redshift in near real time when the new files arrive in the S3 bucket. Which solution will meet these requirements?
A company stores operational data in Amazon S3 using the S3 Standard storage class. Data usage analysis shows that most files are frequently accessed for the first 6 months, occasionally accessed for the next 18 months, and rarely accessed thereafter. The company seeks a cost-effective storage strategy that maintains high availability throughout the data lifecycle.
Which S3 Lifecycle policy should the company implement for their data storage needs?
A company wants to migrate an application and an on-premises Apache Kafka server to AWS. Theapplication processes incremental updates that an on-premises Oracle database sends to the Kafka server. Thecompany wants to use the replatform migration strategy instead of the refactor strategy.Which solution will meet these requirements with the LEAST management overhead?
A company uses Amazon S3 as a data lake. The company sets up a data warehouse by using a multi-nodeAmazon Redshift cluster. The company organizes the data files in the data lake based on the data source ofeach data file.The company loads all the data files into one table in the Redshift cluster by using a separate COPY commandfor each data file location. This approach takes a long time to load all the data files into the table. Thecompany must increase the speed of the data ingestion. The company does not want to increase the cost of theprocess.Which solution will meet these requirements?
A company has three subsidiaries. Each subsidiary uses a different data warehousing solution. The firstsubsidiary hosts its data warehouse in Amazon Redshift. The second subsidiary uses Teradata Vantage onAWS. The third subsidiary uses Google BigQuery.The company wants to aggregate all the data into a central Amazon S3 data lake. The company wants to useApache Iceberg as the table format.A data engineer needs to build a new pipeline to connect to all the data sources, run transformations by usingeach source engine, join the data, and write the data to Iceberg.Which solution will meet these requirements with the LEAST operational effort?
© Copyrights DumpsCertify 2026. All Rights Reserved
We use cookies to ensure your best experience. So we hope you are happy to receive all cookies on the DumpsCertify.