week 6 - language Flashcards

(38 cards)

1
Q

Computational Linguistics

A

The science of how language works using math + logic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

NLP (Natural Language Processing)

A

AI tools for working with language (like translation, chatbots)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Chomsky’s Generative Grammar

A

Early formal models of language structure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

eliza

A

One of the first chatbot programs; mimicked a psychotherapist but lacked true understanding.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Neural Networks

A

AI models inspired by how neurons work.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Word Embeddings

A

words represented as vectors are positioned in a multidimensional space

Examples: “dog” and “cat” vectors placed closely together.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

transformers

A

sequence to sequence model based on deep neural networks (deep learning) with multi head attention

ie, used in translation

input (encoder): English
output (decoder): french

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Self-attention mechanism:

A

Model looks at all words in a sequence, not just the last ones.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

The Transformer consists of two main parts:

A

encoder
embedding
decoder

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

encoder

A

takes the input sequence and maps it into an embedding

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

embedding

A

a n-dimensional vector representing the sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

decoder

A

takes the embedding and turns it into the output sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

self attention

A

allows the transformer to look at the other positions in the input sequence for clues that can help lead to a better encoding of the world

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

self attention diff than previous models

A

can attend at all the words in the sequence, not just the last ones

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Applications of Transformers:

A

Machine translation

Summarization

doc generation

Named Entity Recognition (NER)

Biological sequence analysis

Computer vision

Protein folding

Code generation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

language model

A

Predicts the probability of the next word in a sequence.

17
Q

language model ex

A

GPT, PaLM, LLaMA, Bard, Claude

18
Q

language model training

A

Pre-trained on massive datasets (hundreds of billions of tokens).

19
Q

GPT-style Models

A

Decoder-only models.

Predict one word at a time.

Example: ChatGPT (Generative Pre-trained Transformer).

20
Q

Prompt Engineering

A

Crafting prompts to guide LLM behavior.

21
Q

RLHF (Reinforcement Learning with Human Feedback)

A

Fine-tuning models using human evaluations.

22
Q

in context learning

A

an example to teach the llm how to respond (one shot)

art of asking the right question to get the best output from an llm - enables direct interaction with the llm using only plain lang prompts

23
Q

llm problems

A

performance disparities
bias
misinformation
privacy and security
ethical concerns
environmental concerns

24
Q

toxicity

A

anything that is rude

chatbox could reply with a toxic response

25
language shift
main driver of lanuage endangerment and extinction speakers switch from a native to dominant national language driven by economical factors and marginization of indigenous communities
26
role of ai in language decline
reinforcing english
27
why is english dominant language
global reach data availability research focus
28
BERT
computer model that helps machines understand the meaning of words in a sentence by looking at the words before and after them. It’s used to improve tasks like answering questions, translating language, and understanding text better.
29
What does NLP (Natural Language Processing) focus on? A) Building theoretical models of language B) Practical applications like translation and chatbots C) Creating hardware for AI D) Writing poetry
b
30
What key paper introduced the Transformer model? A) "Attention is All You Need" B) "Generative Grammars" C) "The History Manifesto" D) "Computational Linguistics"
a
31
Which of the following best describes word embeddings? A) Words as vectors based on their context B) Hard-coded grammar rules C) A speech recognition method D) Simple keyword matching
a
32
What is a Language Model’s primary function? A) To summarize texts B) To translate languages C) To predict the next word in a sequence D) To scan images
c
33
What does RLHF stand for in AI training? A) Reinforcement Learning with Human Feedback B) Recursive Language Handling Framework C) Real-time Language Heuristic Function D) Random Learning High Fidelity
a
34
. Which of the following is NOT a risk associated with LLMs? A) Bias and unfairness B) Gender stereotypes C) Toxicity D) Unlimited computing power
d
35
What does ‘Prompt Engineering’ involve? A) Writing computer code to train LLMs B) Creating the best instructions or questions to guide AI responses C) Designing hardware for AI systems D) Debugging AI software
b
36
What is ‘Emergent Abilities’ in Large Language Models? A) Abilities that appear only after specific training for each task B) Abilities that develop spontaneously after a certain model size C) A type of bug in AI models D) Manual programming of new skills
b
37
. Which of these is a core component of the Transformer architecture? A) Recurrent loops B) Self-attention mechanism C) Decision trees D) Support vector machines
b
38
Why is the environmental impact of LLMs a concern? A) They require a lot of data storage B) Training and running models use high energy and produce carbon emissions C) They cannot run on solar power D) They produce toxic output
b