3 aspects of processing power
Computational Capacity
The ability of the hardware to carry out a large number of complex instructions quickly. This is measured in FLOPS (floating-point operations per seconds)
Memory resources
How available is sufficient RAM and VRAM for handling running large models and large volumes of data quickly. Furthermore, these resources ensure for smooth processing and quick access to necessary information.
Efficiency & Speed
The capability to manage high throughput and low latency while optimizing for low energy consumption.
Throughput
The number of data batches are model can process per unit of time.
Preprocessing
One of main tasks of a NN.
Preparing raw data for training to be used by LLM be cleaning and transforming it into a suitable format.
Training the model
One of main tasks of a NN.
Teaching the LLM to understand and generate human-like text by optimizing its parameters using a large dataset.
Deploying the model
One of main tasks of a NN.
Making the trained LLM available for use.
Pre-processing Process
Bag-of-Words Algorithm
Type of pre-processing involving 3 steps:
Tokenization - Text is split into individual words (tokens), often removing punctuation and common “stop words” (such as “and”, “the”, etc.)
Vocabulary Creation - A collection all of the unique words in the text (known as corpus) is created, with each word assigned a unique index.
Vectorization - Each document is represented as a vector of word counts, where the vector length equals the size of the vocabulary, and each element corresponds to the count of a specific word in the document.
X & Y In Bag of words
input and output
4 advantages of Bag-of-Words
Bag-of-Words 4 disadvantages
Advantages of GPUs for being using for LLMs
Tensor Processing Units (TPU)
This is a custom-designed application-specific integrated circuits (ASICs) developed by Google specifically to accelerate machine learning workloads, particularly deep learning tasks.
2 characteristics of TPU
Each TPU unit has 8 cores
Each core has between 8 and 32GB of RAM associated with it
3 Advantages of TPUs being utilized for machine learning
Inference
This refers to the process of using a trained model to make predictions on new, unseen data.
3 usages of TPUs
Clustering with LLMs, 4 advantages
Clustering with LLMs 4 Disadvantages
How does complexity of the model and the dataset in training affect processing power
How does hardware utilization and model architecture in training affect processing power?
How inference latency and throughput as part of the deployment process is affected by processing power?