T-Forms(post E3 #1) Flashcards

1
Q

what is transforms.conf

A

where we specify transformations and lookups that can then be applied to any event.

These transforms and lookups are referenced by name in props.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

main difference between props and transforms

A

-props is responsible for how data breaks and parses

-transforms is responsible for how data looks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Props and Transfroms work together

A

Transforms will not work without props
-props says “do this, but go see transforms for instruction”
-transforms says, “this is how you do it”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Name more functions of Transforms.conf

A

-Manipulates data before it gets indexed
-Transforms and alters raw data
-Has the ability to change metadata

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

List the common uses of Transfroms

A

-Sending events to Null queue
-Separating a single sourcetype into multiple sourcetypes
-Host and source overrides based on Regex
-Delimiter-based field extractions
-Anonymizing data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is a splunk pipeline

A

In Splunk, a pipeline refers to the sequence of data processing stages that data goes through as it is ingested, indexed, and made available for searching and analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is a splunk queue

A

queues are designed to help manage data flow, ensure data integrity, and handle system load efficiently.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is Splunkd considered and what are its subprocesses?

A

*Considered a main process
*the core Splunk process that runs on all Splunk components

1-Parsing queue/parsing pipeline
(linebreaking occurs)
2-Aggregation queue/merging pipeline
(Line merging & Time extraction)
3-Typing queue/typing pipeline
(More Regex occurs here than in parsing pipeline)
4-Indexing queue/indexing pipeline
(syslog out,tcp out, indexer)Final stop is disk after this

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the solution for when you do not want to index unwanted data?

A

Send events to null queue

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Why do we want to prevent unwanted data from indexing ?

A

Do not want extra work, takes more time processing, and is a waste of storage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

List all 4 Splunk pipelines in order

A
  1. Parsing pipeline
  2. Merging pipeline
  3. Typing pipeline
  4. Index pipeline
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Explain process of Splunk pipeline and queues

A

see class notes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What should you do if backlog is caused because process is frozen?

A

Restart Instance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Facts about Regex in Splunk pipeline

A

-More Regex happening in Typing pipeline than parsing pipeline
-More Regex in transfroms.conf(mostly in typing pipeline)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What configurations do we control in Splunk pipeline?

A

props.conf and transforms.conf

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what is a debug event in splunk

A

a log entry or event that contains information about the internal workings, troubleshooting details, or debugging information related to the Splunk platform itself.

17
Q

how do you generate a debug event in splunk

A

command-line tool called logger -located in the bin directory

use it to log custom messages or debug information.

18
Q

Define transforms.conf stanza

A

-DEST_KEY= destination key telling splunk what you want to send to queue

  • FORMAT = specify which queue

-REGEX=DEBUG is looking for any event associated with that match

19
Q

Filepath for transfroms.conf and props.conf

A

Forwarder Level- $SPLUNK_HOME/etc/deployment-apps/app-name/local/props.conf

Indexer Level
$SPLUNK_HOME/etc/master-apps/app-name/local/props.conf

20
Q

Why would we want to split a single sourcetype into multiples ?

A

-a log might contain events that are kind of different from each other

-sometimes logs are just written in a messed up and funny way

21
Q

Explain the process of splitting single sourcetype into multiples

A
  1. First make a reference in props.conf
  2. Then, define in transforms.conf
    (See class notes for pic)

**Same exact steps for Host and source override

22
Q

What is a delimiter? and What is a field extraction?

A

Delimiter=limit
Field Extraction=key value pair

Delimiter-based field extractions in Splunk are used to extract fields from data that is separated by delimiters. Delimiters are characters that are used to separate fields in data, such as commas, spaces, or tabs.

23
Q

Where does transforms.conf go on the searchhead level?

A

$SPLUNK_HOME/etc/apps/app-name/local/transforms.conf

OR

$SPLUNK_HOME/etc/shcluster/apps/<app>/local/transforms.conf</app>

24
Q

What must you use when setting delimiter-based field extractions ?

A

REPORT instead of Transforms.conf

25
Q

Types of data that should be anonymized

A

Title 13 Data = Class of protected data containing private business or personal information

PII = Personal Identifying Information

PHI = Protected Health Information

Title 26 Data = Class protected data containing Federal Tax Information

26
Q

Review how to set up anonymization

A

Look at class notes

27
Q

At what point or stage is anonymizing done?

A

It is done before the indexing pipeline where data is written to disk –i.e. private information is never stored in Splunk.

Anonymizing happens in the typing pipeline.

$SPLUNK_HOME/etc/master-apps/<same_app_name_>/local/props.conf</same_app_name_>

$SPLUNK_HOME/etc/master-apps/<same_app_name_>/local/transforms.conf</same_app_name_>

28
Q

Where else can anonymizing occur?

A

Heavy Forwarder

Clients of the DS:
$SPLUNK_HOME/etc/deployment-apps/<same_app_name_>/local/props.conf</same_app_name_>

$SPLUNK_HOME/etc/deployment-apps/<same_app_name_>/local/transforms.conf</same_app_name_>

Standalone HFs:
$SPLUNK_HOME/etc/apps/<same_app_name_>/local/props.conf</same_app_name_>

$SPLUNK_HOME/etc/apps/<same_app_name_>/local/transforms.conf</same_app_name_>

29
Q

What do you use to hash data?

A

props.conf only + SEDCMD

30
Q

Explain how to hash data with props.conf only + SEDCMD

A

See class notes

31
Q

What should you do when your Splunk pipeline has a fill ratio of 100%?

A

When you have a fill ratio of 100% and a backlog you can open up a second pipeline and direct overfill of events to it

Set number of events a queue can hold = Fill Ratio