Sharing Our Internal Training Material: LLM Terminology Cheat Sheet!
Title: Introducing Our Internal LLM Terminology Reference Guide: A Comprehensive Cheat Sheet for AI Practitioners
In the rapidly evolving field of large language models (LLMs), staying aligned on terminology and core concepts is essential for effective collaboration, research, and implementation. Recognizing this need, our team developed an internal reference resource designed to clarify and unify understanding across various discussions, papers, and evaluations related to AI and LLMs.
Today, we are pleased to share this resource publicly in hopes that it may prove valuable to the broader AI community. The comprehensive cheat sheet offers a structured overview of fundamental topics, making it easier for practitioners to navigate the complex landscape of modern language models.
Overview of the Cheat Sheet
The document is organized into key thematic sections:
- Model Architectures
- Transformer: The foundational architecture powering most modern LLMs.
- Encoder–Decoder: Used in models tailored for tasks like translation.
- Decoder-Only: Models optimized for generative tasks.
-
Mixture of Experts (MoE): Architectures that leverage sparse routing for efficiency.
-
Core Mechanisms
- Attention: The mechanism enabling models to weigh different parts of input data.
- Embeddings: Techniques for representing words and tokens in vector space.
- Quantization: Methods for reducing model size and computational load.
-
Low-Rank Adaptation (LoRA): Techniques for efficient fine-tuning.
-
Training Methodologies
- Pre-training: Initial large-scale training on vast datasets.
- Reinforcement Learning with Human Feedback (RLHF) / RLAIF: Fine-tuning approaches incorporating human preferences.
- Quantized Low-Rank Adaptation (QLoRA): Combining quantization with LoRA for resource-efficient training.
-
Instruction Tuning: Fine-tuning models based on task-specific instructions.
-
Evaluation Benchmarks
- GLUE: A suite for evaluating language understanding.
- MMLU: A benchmark for assessing multilingual and multi-task understanding.
- HumanEval: Measuring code generation quality.
- GSM8K: A benchmark for mathematical problem solving.
Purpose and Audience
This resource is tailored for AI practitioners, researchers, and developers who frequently encounter inconsistent or scattered terminology across LLM literature and documentation. By providing clear, concise definitions and categorizations, we hope to facilitate better communication and understanding within the community.
We believe that shared knowledge accelerates innovation and reduces misunderstandings, paving the way for
Post Comment