ConceptDescription
AGIArtificial General Intelligence (AGI), the point at which AI matches or exceeds the intelligence of humans.
Generative AIAI systems that create new content rather than just analyzing existing data.
Foundation ModelsLarge pre-trained models that serve as the base for various applications.
ArchitectureStructural design of the model. Most modern LLMs use Transformer architectures with attention mechanisms.
Attention MechanismsComponents allowing models to weigh importance of different words when generating text.
TokensBasic units LLMs process; can be words, parts of words, or characters.
TokenizationThe process of breaking text into tokens.
ParametersLearnable weights in the neural network that determine model capabilities. More parameters (measured in billions) generally mean more knowledge and abilities.
Context WindowMaximum amount of text (measured in tokens) an LLM can consider at once.
.safetensorsSecure file format for storing model weights that prevents arbitrary code execution during loading.
Completion/ResponseText generated by the LLM in response to a prompt.
TemperatureSetting that controls randomness in responses—higher values produce more creative outputs.
PromptInput text given to an LLM to elicit a response.
Prompt EngineeringSkill of crafting effective prompts to get desired results from LLMs.
Few-shot LearningProviding examples within a prompt to guide the model toward specific response formats.
Instruction TuningTraining models to follow specific instructions rather than just predicting next words.
HallucinationHallucination in Large Language Models (LLMs) refers to when the model generates false, misleading, or non-factual information that sounds plausible but is incorrect.
EmbeddingsVector representations of words/text that capture semantic meaning and relationships.
RAG (Retrieval-Augmented Generation)Enhancing LLM responses by retrieving relevant information from external sources.
TrainingThe process of teaching an AI model by feeding it data and adjusting its parameters.
InferenceProcess of generating text from the model (as opposed to training).
Fine-tuningProcess of adapting pre-trained models to specific tasks using additional training data.
RLHF (Reinforcement Learning from Human Feedback)Training technique to align LLMs with human preferences and improve safety.
EpochThe number of times a model training process looked through a full data set of images. E.g. The 5th Epoch of a Checkpoint model looked five times through the same data set of images.
float16Half Precision, 16-bit
float32Full Precision, 32-bit