Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Exam Professional-Data-Engineer Topic 3 Question 97 Discussion

Actual exam question for Google's Google Cloud Certified Professional Data Engineer exam
Question #: 97
Topic #: 3
[All Google Cloud Certified Professional Data Engineer Questions]

A web server sends click events to a Pub/Sub topic as messages. The web server includes an event Timestamp attribute in the messages, which is the time when the click occurred. You have a Dataflow streaming job that reads from this Pub/Sub topic through a subscription, applies some transformations, and writes the result to another Pub/Sub topic for use by the advertising department. The advertising department needs to receive each message within 30 seconds of the corresponding click occurrence, but they report receiving the messages late. Your Dataflow job's system lag is about 5 seconds, and the data freshness is about 40 seconds. Inspecting a few messages show no more than 1 second lag between their event Timestamp and publish Time. What is the problem and what should you do?

Show Suggested Answer Hide Answer
Suggested Answer: B

To ensure that the advertising department receives messages within 30 seconds of the click occurrence, and given the current system lag and data freshness metrics, the issue likely lies in the processing capacity of the Dataflow job. Here's why option B is the best choice:

System Lag and Data Freshness:

The system lag of 5 seconds indicates that Dataflow itself is processing messages relatively quickly.

However, the data freshness of 40 seconds suggests a significant delay before processing begins, indicating a backlog.

Backlog in Pub/Sub Subscription:

A backlog occurs when the rate of incoming messages exceeds the rate at which the Dataflow job can process them, causing delays.

Optimizing the Dataflow Job:

To handle the incoming message rate, the Dataflow job needs to be optimized or scaled up by increasing the number of workers, ensuring it can keep up with the message inflow.

Steps to Implement:

Analyze the Dataflow Job:

Inspect the Dataflow job metrics to identify bottlenecks and inefficiencies.

Optimize Processing Logic:

Optimize the transformations and operations within the Dataflow pipeline to improve processing efficiency.

Increase Number of Workers:

Scale the Dataflow job by increasing the number of workers to handle the higher load, reducing the backlog.


Dataflow Monitoring

Scaling Dataflow Jobs

Contribute your Thoughts:

Mattie
1 days ago
Gotta love these Pub/Sub questions. They always seem to have a twist, don't they? I'm going with Option D. Optimize that Dataflow job, baby!
upvoted 0 times
...
Bettina
4 days ago
Option B seems like the way to go. The Dataflow job can't keep up with the backlog in the Pub/Sub subscription. Time to scale up those workers!
upvoted 0 times
...
Nickole
6 days ago
Hmm, I'm not sure the advertising department is the issue here. It sounds like the Dataflow job is the bottleneck. I'd go with Option D and optimize the job.
upvoted 0 times
...
Nikita
8 days ago
But what if the advertising department is causing delays when consuming the messages? Shouldn't we work with them to fix this?
upvoted 0 times
...
Lino
11 days ago
The problem is clearly with the Dataflow job. It's taking too long to process the messages, which is causing the delay for the advertising department. Option D is the correct answer.
upvoted 0 times
...
Paulene
13 days ago
I agree with Sabra. We should optimize the job or increase the number of workers to speed up processing.
upvoted 0 times
...
Sabra
16 days ago
I think the issue might be with the Dataflow job processing the messages too slowly.
upvoted 0 times
...

Save Cancel
az-700  pass4success  az-104  200-301  200-201  cissp  350-401  350-201  350-501  350-601  350-801  350-901  az-720  az-305  pl-300  

Warning: Cannot modify header information - headers already sent by (output started at /pass.php:70) in /pass.php on line 77