Lecture 3 Flashcards
(37 cards)
Waarvoor gebruiken streamingdiensten de kijkgedragingen van gebruikers?
Ze vergelijken jouw kijkgedrag met dat van andere gebruikers om te voorspellen welke shows je leuk vindt.
Welke wiskundige methode ligt ten grondslag aan het aanbevelingssysteem bij streamingdiensten?
Lineaire logistische regressie.
Hoe werkt lineaire logistische regressie in het context van aanbevelingen?
Het model voorspelt de waarschijnlijkheid dat je een bepaalde show leuk vindt op basis van gelijkenissen in kijkgedrag.
Wat is het verschil tussen sceptici en enthousiastelingen als het gaat om AI?
Sceptici focussen vaak op de basiswiskunde en bouwstenen van AI, terwijl enthousiastelingen zich vooral richten op de coole eindresultaten van AI.
Wat is het mechanisme achter veel van de coole dingen die AI kan doen?
Het analyseren van data (zoals gebruikersgedrag) en voorspellen op basis van statistische modellen zoals lineaire logistische regressie.
What is a limitation of the linear/logistic regression model in prediction?
It struggles with situations where the predictor needs to capture “either/or” logic, like deciding if a movie is good based on whether either Meryl Streep or The Rock stars in it.
Give an example illustrating the limitation of logistic regression with the “either/or” problem.
If a movie has Meryl Streep or The Rock, you predict it’s good. But if it has both, logistic regression struggles because they excel in different genres, so combining them confuses the model.
What is the “XOR problem” in AI?
It’s a classic problem where AI models like logistic regression cannot learn the logical function of “this or that” (exclusive OR).
Why do we need neural networks to solve the XOR problem?
Because neural networks can perform multiple logistic regressions in parallel and combine their results to capture complex logical relationships.
How do neural networks improve on simple logistic regression?
Neural networks build intermediate predictors by combining logistic regressions in layers, which extract useful combinations of raw data and improve understanding.
What is the role of intermediate predictors in neural networks?
They create more informative features from the original data, enabling the AI to make more accurate predictions by considering more complex patterns.
What is the “Universal Approximation Theorem”?
It states that neural networks can approximate any relationship between inputs and outputs, meaning they can model almost any logical or functional pattern.
Why can’t simple logistic regression handle XOR but neural networks can?
Because logistic regression only models linear relationships, while neural networks combine many logistic regressions in layers, allowing them to model non-linear, complex relationships like XOR.
Summarize why neural networks are a much better way to predict data compared to simple logistic regression
Neural networks create multiple intermediate steps that transform raw data into meaningful features, enabling them to capture complex patterns and logic that logistic regression cannot
Does ChatGPT work exactly like a neural network?
Not exactly. It uses neural networks as building blocks within a broader architecture called a transformer model
What are language models like ChatGPT trained to do?
Predict the next word in a sequence—not the correct answer, just the most likely one based on previous words.
Why is predicting the next word difficult for AI?
While it’s a basic task, doing it well requires a lot of background knowledge, context awareness, and handling of word order.
Why don’t regular neural networks work well for language tasks?
Because they don’t consider word order, which is crucial in understanding language.
What model solves this problem of word order and meaning?
The transformer model.
What does the encoder part of the transformer do?
Turns words into numbers that best capture the meaning of the input.
What are embeddings?
Numerical representations of words based on their co-occurrence with other words.
How are embeddings created in basic neural networks?
By trying to predict the next word and observing which words occur together frequently.
What is positional encoding in transformers?
A numerical representation added to embeddings to capture a word’s position in the sentence.
What happens when you combine embedding and positional encoding?
You get a richer numerical representation that includes both meaning and position (e.g., won = {0.4, 0.4, 0} + {1, 0.9, -0.8} = {1.4, 1.3, 0.8}).