Commonly Asked Demo Questions Flashcards

1
Q

How big is the Datadog agent? What is the agent overhead?

A

In terms of resource consumption the Datadog agent consumes roughly:
Resident memory (actual RAM used): 50MB
CPU Runtime: less than 1% of averaged runtime
Disk: (Linux 120MB, Windows: 60MB)
Network: 10-50 KB of bandwidth per minute
*The stats listed above are based on an EC2 m1.large instance running for 10+ days.

For APM, resource consumption can vary a lot depending on actual usage. For minimal usage, the order of magnitude is similar to the numbers listed above.

Max total trace size: 10MB

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Objection: But you don’t support on-prem!

A

“Actually, it’s a common misconception that we don’t support on-premise use cases. Many of our customers, including one of the largest healthcare platforms, used Datadog to monitor their migration from a mostly on-prem environment to the cloud.”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Do we have to use the agent?

A

The agent is not necessary, but, it is highly recommended
For starters, the OS metrics that you get are a very helpful baseline for your servers.
You also get more metrics (versus Cloudwatch for example)
Also, some of our integrations require the agent to collect metrics.
To enable APM, the agent + client (this refers to the programming language - python, ruby, go) is required (Linux)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How do you install the agent with Chef/Puppet?

A

We have a cookbook for Chef and a module for Puppet

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How does Datadog monitor containers?

A

We monitor containers no matter how frequently they are created or destroyed. We also monitor orchestration tools like Kubernetes, Mesos, or Amazon ECS
We have what we call, Autodiscovery - which auto-discovers containers and apps running on them. So we can automatically detect and monitor any changes to your containerized environment and always have visibility into the apps running on them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Can you run a version of Datadog on-premise?

A

The agent is what you would install on-premise to collect metrics/events from your local environment and send it to your Datadog account

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

I’m concerned about security issues related to sending data to an external 3rd party

A

The data collected by default is very benign, and they can audit our Agent’s code via our Github account to verify everything that it does.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

I’m concerned about security issues related to having more places of connections from my infrastructure to the internet.

A

You can set up the agent to report metrics via Proxy to reduce the number of hosts that actually have a connection to the internet. Our backend hosts all the data and the DataDog UI queries it.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do you transmit information over the internet?

A

All data transmitted between Datadog and Datadog users is protected using Transport Layer Security (TLS) and HTTP Strict Transport Security (HSTS).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How is data encrypted in-flight/at rest?

A

Data is encrypted at rest using secure symmetric cipher - AES.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How is the data stored?

A

We’re stored entirely on AWS, and rely on their security for storage. If they ask if we’re multi-region - the answer is no. We’re on East.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What security certifications do you have?

A

CSA STAR and AICPA SOC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What metrics are collected for each integration? How are they decided?

A

Our development team undergoes an analysis of which metrics are noteworthy to collect, and then programs them into each integration. We list which metrics are being collected for each integration on our Docs page.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How can we get custom metrics in?

A

We have a StatsD handler called DogStatsD as part of the agent if you are using StatsD. We also have a REST-ful API

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How many custom metrics can you take?

A

We can take an unlimited number of custom metrics. Included in our Pro pricing is 100 custom metrics per host. If you will have more, we work out pricing based on the additional volume. However, we are able to cut down typically on large amounts of custom metrics by combining metrics with tags. Pricing for Lambda falls under custom metrics pricing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Can you collect and show logs in Datadog?

A

Yes, Datadog’s log management tool can also show logs from other integrated tools in the event stream, or overlay them onto graphs with metrics.

17
Q

Does DataDog integrate with Splunk?

A

Yes. However, logs from other sources are shown in the ‘Event Stream,’ and are not available in the ‘Log Explorer.’ The Log Explorer has many features for searching, visualization, and correlation with infrastructure or APM that aren’t available with logs that come from other sources.

18
Q

How is the data parsed from logs determined to be a metric or event?

A

If you’re parsing integer/float values, it’s a metric. If it’s text, it’s an event. Metrics are considered custom when parsing your own logs from other log management tools.

19
Q

How are you different from AWS CloudWatch?

A

Datadog gets much more granular (down to 1 second) and hold that granularity for the full retention period. Cloudwatch rolls up to 1 hour for 15 months. They start out at 1 min granularity and roll up to 5 mins after 15 days.

We are also much more robust for alerting and graphing based on our ability to aggregate data from multiple servers, and the ability to “slice and dice” this data based on our tagging.

Datadog provides 75-120 system level metrics and Cloudwatch is around 13 metrics

Lastly, AWS CloudWatch is only one source for us. We will take multiple sources of data and allow for correlation analysis between metrics and events.

20
Q

What is the flamegraph?

A

The flamegraph in Datadog represents traces. It represents one request and is made up of spans. Spans represent the different services involved in that request

21
Q

What is a trace?

A

A trace is a visualization of the communication across different services to process a request and the time spent by each service

22
Q

What is the data retention time frame? Can it be extended?

A

We retain data at full granularity for 15 months (Infrastructure and APM metrics), but can easily do more as needed.

23
Q

How do you sample traces?

A

We sample fast (performing well) and slow (not performing well) traces using signature-based sampling, which also removes traces that are too similar to each other

24
Q

What do custom metrics look like inside of Datadog?

A

They look the exact same as all of the other metrics. If you can “increment it”, you can send it to Datadog. Metric = value + timestamp so it can be plotted

25
Q

Can you send alerts via email, SMS, ticketing systems?

A

Yes. Alerts can be sent via email, or, to alerting systems like PagerDuty, OpsGenie and VictorOps. Alerts can also be sent via SMS and to ticketing systems like JIRA or Zendesk using our integrations

26
Q

How often does Datadog collect data?

A

Datadog is collecting data at 10 to 20 second intervals from the agent, so about 15 seconds
The agent’s collector itself, runs every check sequentially and then sleeps for 15 seconds, so depending on the duration of every check it can run every 15 seconds or more.
DogStatsD, on the other hand works differently, DogStatsD receives metrics (instead of collecting them like the collector does), and then flushes them every 10 seconds.

27
Q

What’s the level of granularity that data is stored?

A

We’re capable of storing data at 1 second, although the default is 15 seconds.

28
Q

What languages does Datadog support?

A

Ruby, Java, Python, Go, C#/.NET, plus others through DogStatsD & API

29
Q

Is the product API enabled?

A

Mostly everything in the UI is API enabled

30
Q

What languages are supported for APM?

A

Java, Python, Ruby, Go, Node.js and popular frameworks/libraries for those languages

31
Q

What Database/other integrations are supported for APM?

A

It depends on the language and we’re always releasing new integrations. Which integrations are we looking for?

32
Q

Do you have “fill-in-the-blank” integration?

A

We’re always releasing new integrations as well as offer community contributed integrations for tools we haven’t built yet. Let’s pull up the most up-to-date list. Which integrations are we looking for?

33
Q

What is “Event metadata”?

A

Event metadata is data that provides information about the event’s main data, such as: any tags associated with that event, the integration the event came from, the time and date of creation and the user or creator of the event. The metadata we collect for each event depends on what’s available from the integration.

34
Q

What are tags?

A

Tags are labels applied to metrics and hosts and can be used for filtering in graphs and alerts. They are key: value pairs