Spring Sale 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: save70

Free Access Snowflake ARA-C01 New Release

Page: 13 / 14
Total 182 questions

SnowPro Advanced: Architect Certification Exam Questions and Answers

Question 49

An Architect is designing Snowflake architecture to support fast Data Analyst reporting. To optimize costs, the virtual warehouse is configured to auto-suspend after 2 minutes of idle time. Queries are run once in the morning after refresh, but later queries run slowly.

Why is this occurring?

Options:

A.

The warehouse is not large enough.

B.

The warehouse was not configured as a multi-cluster warehouse.

C.

The warehouse was not created with USE_CACHE = TRUE.

D.

When the warehouse was suspended, the cache was dropped.

Question 50

What considerations apply when using database cloning for data lifecycle management in a development environment? (Select TWO).

Options:

A.

Any pipes in the source are not cloned.

B.

Any pipes in the source referring to internal stages are not cloned.

C.

Any pipes in the source referring to external stages are not cloned.

D.

The clone inherits all granted privileges of all child objects, including the database.

E.

The clone inherits all granted privileges of all child objects, excluding the database.

Question 51

An Architect wants to stream website logs near real time to Snowflake using the Snowflake Connector for Kafka.

What characteristics should the Architect consider regarding the different ingestion methods? (Select TWO).

Options:

A.

Snowpipe Streaming is the default ingestion method.

B.

Snowpipe Streaming supports schema detection.

C.

Snowpipe has lower latency than Snowpipe Streaming.

D.

Snowpipe Streaming automatically flushes data every one second.

E.

Snowflake can handle jumps or resetting offsets by default.

Question 52

A Snowflake Architect Is working with Data Modelers and Table Designers to draft an ELT framework specifically for data loading using Snowpipe. The Table Designers will add a timestamp column that Inserts the current tlmestamp as the default value as records are loaded into a table. The Intent is to capture the time when each record gets loaded into the table; however, when tested the timestamps are earlier than the loae_take column values returned by the copy_history function or the Copy_HISTORY view (Account Usage).

Why Is this occurring?

Options:

A.

The timestamps are different because there are parameter setup mismatches. The parameters need to be realigned

B.

The Snowflake timezone parameter Is different from the cloud provider's parameters causing the mismatch.

C.

The Table Designer team has not used the localtimestamp or systimestamp functions in the Snowflake copy statement.

D.

The CURRENT_TIMEis evaluated when the load operation is compiled in cloud services rather than when the record is inserted into the table.

Page: 13 / 14
Total 182 questions