linux foundation PCA Exam Questions

Questions for the PCA were updated on : Nov 21 ,2025

Page 1 out of 4. Viewing questions 1-15 out of 60

Question 1

Which function would you use to calculate the 95th percentile latency from histogram data?

  • A. quantile_over_time(0.95, http_request_duration_seconds[5m])
  • B. histogram_quantile(0.95, sum(rate(http_request_duration_seconds_bucket[5m])) by (le))
  • C. percentile(http_request_duration_seconds, 0.95)
  • D. topk(0.95, http_request_duration_seconds)
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
To calculate a percentile (e.g., 95th percentile) from histogram data in Prometheus, the correct
function is histogram_quantile(). It estimates quantiles based on cumulative bucket counts.
Example:
histogram_quantile(0.95, sum(rate(http_request_duration_seconds_bucket[5m])) by (le))
This computes the 95th percentile request duration across all observed instances over the last 5
minutes.
Reference: Prometheus documentation – histogram_quantile() Function, Working with Histogram
Buckets.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 2

Which Prometheus component handles service discovery?

  • A. Alertmanager
  • B. Prometheus Server
  • C. Pushgateway
  • D. Node Exporter
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The Prometheus Server is responsible for service discovery, which identifies the list of targets to
scrape. It integrates with multiple service discovery mechanisms such as Kubernetes, Consul, EC2,
and static configurations.
This allows Prometheus to automatically adapt to dynamic environments without manual
reconfiguration.
Reference: Prometheus documentation – Service Discovery Mechanisms, Prometheus Architecture.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 3

What are the four golden signals of monitoring as defined by Google’s SRE principles?

  • A. Traffic, Errors, Latency, Saturation
  • B. Requests, CPU, Memory, Latency
  • C. Availability, Logging, Errors, Throughput
  • D. Utilization, Load, Disk, Network
Answer:

A

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The Four Golden Signals—Traffic, Errors, Latency, and Saturation—are key service-level indicators
defined by Google’s Site Reliability Engineering (SRE) discipline.
Traffic: Demand placed on the system (e.g., requests per second).
Errors: Rate of failed requests.
Latency: Time taken to serve requests.
Saturation: How “full” the system resources are (CPU, memory, etc.).
Prometheus and its metrics-based model are ideal for capturing these signals.
Reference: Google SRE Book & Prometheus Best Practices – Golden Signals and SLO Monitoring.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 4

Which Alertmanager feature allows you to temporarily stop notifications for a specific alert?

  • A. Silence
  • B. Inhibition
  • C. Deduplication
  • D. Grouping
Answer:

A

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The Silence feature in Alertmanager allows operators to mute specific alerts for a defined period.
Each silence includes a matcher (labels), a creator, a comment, and an expiration time.
Silencing is useful during maintenance windows or known outages to prevent alert noise. Unlike
inhibition, silences are manual and explicit.
Reference: Prometheus documentation – Silences in Alertmanager, Managing Alerts During
Maintenance Windows.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 5

What does the rate() function in PromQL return?

  • A. The total increase of a counter over a range.
  • B. The per-second rate of increase of a counter metric.
  • C. The average of all values in a vector.
  • D. The number of samples in a range vector.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The rate() function calculates the average per-second rate of increase of a counter over the specified
range. It smooths out short-term fluctuations and adjusts for counter resets.
Example:
rate(http_requests_total[5m])
returns the number of requests per second averaged over the last five minutes. This function is
frequently used in dashboards and alerting expressions.
Reference: Prometheus documentation – rate() Function Definition, Counter Handling in PromQL.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 6

Where does Prometheus store its time series data by default?

  • A. In-memory only.
  • B. In an embedded TSDB on local disk.
  • C. In an external database such as InfluxDB.
  • D. In etcd.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
By default, Prometheus stores its time series data in a local, embedded Time Series Database (TSDB)
on disk. The data is organized in block files under the data/ directory inside Prometheus’s storage
path.
Each block typically covers two hours of data, containing chunks, index, and metadata files. Older
blocks are compacted and deleted based on retention settings.
Reference: Prometheus documentation – Storage, TSDB Internals, Retention and Compaction.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 7

Which Alertmanager feature prevents duplicate notifications from being sent?

  • A. Grouping
  • B. Inhibition
  • C. Deduplication
  • D. Silencing
Answer:

C

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
Deduplication in Alertmanager ensures that identical alerts from multiple Prometheus servers or rule
evaluations do not trigger duplicate notifications.
Alertmanager compares alerts based on their labels and fingerprints; if an alert with identical labels
already exists, it merges or refreshes the existing one instead of creating a new notification.
This mechanism is essential in high-availability setups where multiple Prometheus instances monitor
the same targets.
Reference: Prometheus documentation – Alertmanager Architecture, Deduplication, Grouping, and
Routing Logic.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 8

What is the role of the Pushgateway in Prometheus?

  • A. To scrape short-lived targets directly.
  • B. To receive metrics pushed by short-lived batch jobs for later scraping by Prometheus.
  • C. To store metrics long-term for historical analysis.
  • D. To visualize metrics in Grafana.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The Pushgateway is a Prometheus component used to handle short-lived batch jobs that cannot be
scraped directly. These jobs push their metrics to the Pushgateway, which then exposes them for
Prometheus to scrape.
This ensures metrics persist beyond the job’s lifetime. However, it’s not designed for continuously
running services, as metrics in the Pushgateway remain static until replaced.
Reference: Prometheus documentation – Pushgateway Overview, Best Practices for Short-Lived Jobs.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 9

What does the increase() function do in PromQL?

  • A. Calculates the percentage increase of a counter over time.
  • B. Returns the absolute increase in a counter over a specified range.
  • C. Calculates the derivative of a gauge over time.
  • D. Returns the total sum of values in a vector.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The increase() function computes the total increase in a counter metric over a specified range vector.
It accounts for counter resets and only measures the net change in the counter’s value during the
time window.
Example:
increase(http_requests_total[5m])
This query returns how many HTTP requests occurred in the last five minutes. Unlike rate(), which
provides a per-second average rate, increase() gives the absolute number of increments.
Reference: Prometheus documentation – PromQL Function Reference, Counter Handling and
increase() vs. rate().

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 10

What does the evaluation_interval parameter in the Prometheus configuration control?

  • A. How often Prometheus scrapes targets.
  • B. How often Prometheus evaluates recording and alerting rules.
  • C. How often Prometheus compacts the TSDB data blocks.
  • D. How often Prometheus sends metrics to remote storage.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The evaluation_interval parameter defines how frequently Prometheus evaluates its recording and
alerting rules. It determines the schedule at which the rule engine runs, checking whether alert
conditions are met and generating new time series for recording rules.
For example, setting:
global:
evaluation_interval: 30s
means Prometheus evaluates all configured rules every 30 seconds. This setting differs from
scrape_interval, which controls how often Prometheus collects data from targets.
Having a proper evaluation interval ensures alerting latency is balanced with system performance.
Reference: Prometheus documentation – Configuration File Reference, Rule Evaluation Cycle, Global
vs Job-Specific Parameters.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 11

Which field in alerting rules files indicates the time an alert needs to go from pending to firing state?

  • A. duration
  • B. interval
  • C. timeout
  • D. for
Answer:

D

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
In Prometheus alerting rules, the for field specifies how long a condition must remain true
continuously before the alert transitions from the pending to the firing state. This feature prevents
transient spikes or brief metric fluctuations from triggering false alerts.
Example:
alert: HighRequestLatency
expr: http_request_duration_seconds_avg > 1
for: 5m
labels:
severity: warning
annotations:
description: "Request latency is above 1s for more than 5 minutes."
In this configuration, Prometheus evaluates the expression every rule evaluation cycle. The alert only
fires if the condition (http_request_duration_seconds_avg > 1) remains true for 5 consecutive
minutes. If it returns to normal before that duration, the alert resets and never fires.
This mechanism adds stability and noise reduction to alerting systems by ensuring only sustained
issues generate notifications.
Reference:
Verified from Prometheus documentation – Alerting Rules Configuration Syntax, Pending vs. Firing
States, and Best Practices for Alert Timing and Thresholds sections.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 12

Which of the following signal belongs to symptom-based alerting?

  • A. Disk space
  • B. CPU usage
  • C. Database memory utilization
  • D. API latency
Answer:

D

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
Symptom-based alerting focuses on user-visible problems or service-impacting symptoms rather
than low-level resource metrics. In Prometheus and Site Reliability Engineering (SRE) practices, alerts
should signal conditions that affect users’ experience — such as high latency, request failures, or
service unavailability — instead of merely reflecting internal resource states.
Among the options, API latency directly represents the performance perceived by end users. If API
response times increase, it immediately impacts user satisfaction and indicates a possible service
degradation.
In contrast, metrics like disk space, CPU usage, or database memory utilization are cause-based
metrics — they may correlate with problems but do not always translate into observable user
impact.
Prometheus alerting best practices recommend alerting on symptoms (via RED metrics — Rate,
Errors, Duration) while using cause-based metrics for deeper investigation and diagnosis, not for
immediate paging alerts.
Reference:
Verified from Prometheus documentation – Alerting Best Practices, Symptom vs. Cause Alerting, and
RED/USE Monitoring Principles sections.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 13

What popular open-source project is commonly used to visualize Prometheus data?

  • A. Kibana
  • B. Grafana
  • C. Thanos
  • D. Loki
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The most widely used open-source visualization and dashboarding platform for Prometheus data is
Grafana. Grafana provides native integration with Prometheus as a data source, allowing users to
create real-time, interactive dashboards using PromQL queries.
Grafana supports advanced visualization panels (graphs, heatmaps, gauges, tables, etc.) and enables
users to design custom dashboards to monitor infrastructure, application performance, and service-
level objectives (SLOs). It also provides alerting capabilities that can complement or extend
Prometheus’s own alerting system.
While Kibana is part of the Elastic Stack and focuses on log analytics, Thanos extends Prometheus for
long-term storage and high availability, and Loki is a log aggregation system. None of these tools
serve as the primary dashboarding solution for Prometheus metrics the way Grafana does.
Grafana’s seamless Prometheus integration and templating support make it the de facto standard
visualization tool in the Prometheus ecosystem.
Reference:
Verified from Prometheus documentation – Visualizing Data with Grafana, and Grafana
documentation – Prometheus Data Source Integration and Dashboard Creation Guide.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 14

When can you use the Grafana Heatmap panel?

  • A. You can use it to graph a counter metric.
  • B. You can use it to graph a histogram metric.
  • C. You can use it to graph an info metric.
  • D. You can use it to graph a gauge metric.
Answer:

B

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
The Grafana Heatmap panel is best suited for visualizing histogram metrics collected from
Prometheus. Histograms provide bucketed data distributions (e.g., request durations, response
sizes), and the heatmap effectively displays these as a two-dimensional density chart over time.
In Prometheus, histogram metrics are exposed as multiple time series with the _bucket suffix and
the label le (less than or equal). Grafana interprets these buckets to create visual bands showing how
frequently different value ranges occurred.
Counters, gauges, and info metrics do not have bucketed distributions, so a heatmap would not
produce meaningful output for them.
Reference:
Verified from Grafana documentation – Heatmap Panel Overview, Visualizing Prometheus
Histograms, and Prometheus documentation – Understanding Histogram Buckets.

Discussions
vote your answer:
A
B
C
D
0 / 1000

Question 15

What is a rule group?

  • A. It is a set of rules that are executed sequentially.
  • B. It is the set (the group) of all the rules in a file.
  • C. It is a set of rules, split into groups by type.
  • D. It is a set of rules that are grouped by labels.
Answer:

A

User Votes:
A
50%
B
50%
C
50%
D
50%

Explanation:
In Prometheus, a rule group is a logical collection of recording and alerting rules that are evaluated
sequentially at a specified interval. Rule groups are defined in YAML files under the groups: key, with
each group containing a name, an interval, and a list of rules.
For example:
groups:
- name: example
interval: 1m
rules:
- record: job:http_inprogress_requests:sum
expr: sum(http_inprogress_requests) by (job)
All rules in a group share the same evaluation schedule and are executed one after another. This
ensures deterministic order, especially when one rule depends on another’s result.
Reference:
Verified from Prometheus documentation – Rule Configuration, Rule Groups and Evaluation Order,
and Recording & Alerting Rules Guide.

Discussions
vote your answer:
A
B
C
D
0 / 1000
To page 2