BlackFriday 2024! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Amazon Exam DAS-C01 Topic 2 Question 82 Discussion

Actual exam question for Amazon's DAS-C01 exam
Question #: 82
Topic #: 2
[All DAS-C01 Questions]

A company is building an analytical solution that includes Amazon S3 as data lake storage and Amazon Redshift for data warehousing. The company wants to use Amazon Redshift Spectrum to query the data that is stored in Amazon S3.

Which steps should the company take to improve performance when the company uses Amazon Redshift Spectrum to query the S3 data files? (Select THREE )

Use gzip compression with individual file sizes of 1-5 GB

Show Suggested Answer Hide Answer
Suggested Answer: B, C, D

Contribute your Thoughts:

Blair
7 months ago
Good points, everyone. But I'm not sure about the whole 'KB-sized files' thing. That seems a bit overkill, don't you think? 1-5 GB seems like a good sweet spot.
upvoted 0 times
...
Kati
7 months ago
You guys are on the right track. But I also think we should use gzip compression to reduce the file sizes. Smaller files will be faster to scan.
upvoted 0 times
Jolene
7 months ago
C) Split the data into KB-sized files.
upvoted 0 times
...
Dylan
7 months ago
B) Partition the data based on the most common query predicates
upvoted 0 times
...
Yong
7 months ago
A) Use a columnar storage file format
upvoted 0 times
...
...
Aleisha
7 months ago
I agree, the columnar storage format is crucial for efficient querying. And partitioning the data based on the most common predicates will help Redshift Spectrum retrieve the relevant data quickly.
upvoted 0 times
...
Robt
7 months ago
Hmm, this question seems to be testing our understanding of how to optimize performance when using Amazon Redshift Spectrum. I think the key here is to use columnar storage file formats, partition the data, and keep the file sizes consistent.
upvoted 0 times
...

Save Cancel