LLM Flashcards

(23 cards)

1
Q

What is subword tokenization?

A

Words split into meaningful sub-parts

Common in modern tokenizers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the purpose of TrainingArguments() in fine-tuning?

A

Customize training settings

See documentation for all parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What does the output_dir parameter in TrainingArguments represent?

A

Output directory for the fine-tuned model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the significance of num_train_epochs in TrainingArguments?

A

Number of training epochs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does learning_rate control in training?

A

For optimizer’s learning adjustments

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What do per_device_train_batch_size and per_device_eval_batch_size define?

A

The batch size for training and evaluation respectively

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the role of the Trainer class?

A

To manage the training process

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the eval_dataset used for in the Trainer class?

A

The data used for evaluation during training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

In the training output, what does eval_loss indicate?

A

The loss value during evaluation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How are predicted labels derived from model outputs?

A

Using torch.argmax on outputs.logits

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does the label_map dictionary represent?

A

Mapping of predicted labels to sentiment categories

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the function of the pipeline() method?

A

Streamlines tasks with automatic model and tokenizer selection

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is full fine-tuning?

A

Updating the entire model weights

Computationally expensive

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is partial fine-tuning?

A

Only task-specific layers are updated; some layers are fixed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Define transfer learning in the context of fine-tuning.

A

Adapting a pre-trained model to a different but related task

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is zero-shot learning?

A

No examples are provided for the model to learn from

17
Q

Fill in the blank: In one-shot learning, the model is provided with _______.

18
Q

What does load_dataset() do?

A

Loads a dataset for fine-tuning

19
Q

What is the purpose of tokenization in machine learning?

A

Converts text into a format suitable for model training

20
Q

What does the tokenizer function do to the input data?

A

It processes the text data into tensors for model input

21
Q

What is the output of the tokenization process?

A

A dictionary containing input_ids, attention_mask, etc.

22
Q

What does the tokenize_function do?

A

Applies tokenization to text data

23
Q

What is the difference between tokenizing in batches and tokenizing row by row?

A

Batch tokenization processes multiple rows at once, while row by row processes one at a time