week 6 - language Flashcards
(38 cards)
Computational Linguistics
The science of how language works using math + logic
NLP (Natural Language Processing)
AI tools for working with language (like translation, chatbots)
Chomsky’s Generative Grammar
Early formal models of language structure.
eliza
One of the first chatbot programs; mimicked a psychotherapist but lacked true understanding.
Neural Networks
AI models inspired by how neurons work.
Word Embeddings
words represented as vectors are positioned in a multidimensional space
Examples: “dog” and “cat” vectors placed closely together.
transformers
sequence to sequence model based on deep neural networks (deep learning) with multi head attention
ie, used in translation
input (encoder): English
output (decoder): french
Self-attention mechanism:
Model looks at all words in a sequence, not just the last ones.
The Transformer consists of two main parts:
encoder
embedding
decoder
encoder
takes the input sequence and maps it into an embedding
embedding
a n-dimensional vector representing the sequence
decoder
takes the embedding and turns it into the output sequence
self attention
allows the transformer to look at the other positions in the input sequence for clues that can help lead to a better encoding of the world
self attention diff than previous models
can attend at all the words in the sequence, not just the last ones
Applications of Transformers:
Machine translation
Summarization
doc generation
Named Entity Recognition (NER)
Biological sequence analysis
Computer vision
Protein folding
Code generation
language model
Predicts the probability of the next word in a sequence.
language model ex
GPT, PaLM, LLaMA, Bard, Claude
language model training
Pre-trained on massive datasets (hundreds of billions of tokens).
GPT-style Models
Decoder-only models.
Predict one word at a time.
Example: ChatGPT (Generative Pre-trained Transformer).
Prompt Engineering
Crafting prompts to guide LLM behavior.
RLHF (Reinforcement Learning with Human Feedback)
Fine-tuning models using human evaluations.
in context learning
an example to teach the llm how to respond (one shot)
art of asking the right question to get the best output from an llm - enables direct interaction with the llm using only plain lang prompts
llm problems
performance disparities
bias
misinformation
privacy and security
ethical concerns
environmental concerns
toxicity
anything that is rude
chatbox could reply with a toxic response