Skip to main content
BVDNETBVDNET
ServicesWorkLibraryAboutPricingBlogContact
Contact
  1. Home
  2. AI Woordenboek
  3. Models & Architecture

Models & Architecture

9 concepts

All categoriesModels & ArchitectureTools & FrameworksAgentic AISafety & EthicsIndustry & BusinessPractical ApplicationsCore Concepts
LoRA (Low-Rank Adaptation)
Intermediate
Models & Architecture

LoRA (Low-Rank Adaptation)

An efficient fine-tuning method that trains only small adapter layers instead of the full model

What Is Model Distillation? How Knowledge Transfer Makes AI Smaller & Faster
Intermediate
Models & Architecture

Model Distillation

Training a smaller 'student' model to replicate a larger 'teacher' model's capabilities at a fraction of the cost and latency

What Is Perplexity in NLP? The Key Metric for Language Model Evaluation
Intermediate
Models & Architecture

Perplexity in NLP

The standard metric for evaluating language model quality — measuring how well a model predicts text, where lower values indicate better language understanding

Quantization
Intermediate
Models & Architecture

Quantization

Reducing model weight precision from 16/32-bit to 8/4-bit to shrink size and speed up inference

RAG (Retrieval-Augmented Generation)
Intermediate
Models & Architecture

RAG (Retrieval-Augmented Generation)

A technique that combines LLMs with external knowledge retrieval to improve accuracy and reduce hallucinations

RLHF (Reinforcement Learning from Human Feedback)
Advanced
Models & Architecture

RLHF (Reinforcement Learning from Human Feedback)

A training technique that uses human preference ratings to align LLM behavior with human values

Transformer
Intermediate
Models & Architecture

Transformer

The neural network architecture underlying all modern LLMs, using attention mechanisms to process text

What Is the Attention Mechanism? Self-Attention & Multi-Head Attention Explained
Advanced
Models & Architecture

Attention Mechanism

The mathematical mechanism that allows transformers to dynamically focus on the most relevant parts of the input when processing each token

What Is the KV Cache? How Key-Value Caching Accelerates LLM Inference
Advanced
Models & Architecture

KV Cache

A memory optimization that stores previously computed key-value pairs in transformer attention layers — avoiding redundant computation and accelerating generation 3-5×

BVDNETBVDNET

Web development and AI automation. Done properly.

Company

  • About
  • Contact
  • FAQ

Resources

  • Services
  • Work
  • Library
  • Blog
  • Pricing

Connect

  • LinkedIn
  • GitHub
  • Twitter / X
  • Email

© 2026 BVDNET. All rights reserved.

Privacy Policy•Terms of Service•Cookie Policy