Weekend Sale Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Architect Exams Heroku-Architect Syllabus Exam Questions Answers

Salesforce Certified Heroku Architect(SP23) Questions and Answers

Question 21

What are the three access levels a user can have in a Heroku Organization account?

Options:

A.

Jedi. Padawan, and Co-pilot

B.

Owner, Author and Viewer

C.

Admin. Programmer, and Operator

D.

Admin, Member, and Viewer

Question 22

Which three actions can a user perform on an app in a Heroku Enterprise Team, if that user has been granted only the "deploy" permission for that app? Choose 3 answers.

Options:

A.

Provision an add-on with a free plan

B.

Manage the app's config vars.

C.

Transfer the app to another Enterprise Team.

D.

Roll back a release

Question 23

Universal Containers (UC) has a Heroku Postgres database attached to a Heroku app, which is running in the Common Runtime.

What should an Architect recommend UC do to secure their database to only trusted locations?

Options:

A.

Move the database and the app into the Private Space.

B.

Add a Trusted IP range tothe Heroku Postgres configuration.

C.

Set up a VPN from the trusted locations to the app.

D.

Set up a VPC in a one-off dyno when access to Heroku Postgres is needed.

Question 24

Universal Containers (UC)uses Apache Kafka on Heroku to stream shipment inventory data in real time throughout the world. A Kafka topic is used to send messages with updates on the shipping container GPS coordinates as they are in transit. UC is using a Heroku Kafka basic-0 plan.The topic was provisioned with 8 partitions, 1 week of retention, and no compaction. The keys for the events are being assigned by Heroku Kafka, which means that they will be randomly distributed between the partitions.

UC has a single-dyno consumer application that persists the data to their Enterprise Data Warehouse (EDW). Recently, they've been noticing data loss in the EDW.

What should an Architect with Kafka experience recommend?

Options:

A.

Enable compaction on the topic to drop older messages, which will drop older messages with the same key.

B.

Upgrade to a larger Apache Kafka for Heroku plan, which has greater data capacity.

C.

Use Heroku Redis to store message receipt information to account for "at-least" once delivery, which will guarantee that messagesare never processed more than once. Scale up the consumer dynos to match the number of partitions so that there is one process for each partition.