Design & LLM Interaction LLM Reasoning and Architecture Reinforcement Learning for LLMs

Does model access level determine which specialization techniques work?

Different specialization approaches require different levels of access to a model's internals. Understanding this constraint helps practitioners choose realistic techniques for their domain adaptation goals.

Note · 2026-02-21 · sourced from Domain Specialization
How do you build domain expertise into general AI models? How should researchers navigate LLM reasoning research?

The domain specialization survey organizes the technique landscape around a single governing variable: how much access does the practitioner have to the model's internals? This produces three tiers that are not just organizational — they determine the ceiling on what specialization can achieve.

Black-box (external augmentation): No access to model parameters, gradients, or loss values. Techniques: RAG, tool use, output post-processing, prompt injection. Domain knowledge is incorporated into the input or used to filter the output. The model itself is unchanged. This is the most accessible tier — any API user can apply it — but the specialization is shallow: the model applies pre-existing general capabilities to domain-enriched prompts. Knowledge that isn't explicitly in the context window cannot be activated.

Grey-box (prompt crafting): Access to gradient or loss values, allowing finer control over model behavior without modifying parameters. Techniques: continuous prompt tuning, soft prompts, learnable prompt vectors. The model's behavior is shaped by optimized prompt representations rather than natural language instructions. More powerful than discrete prompting because the optimization happens in embedding space rather than token space, but still does not change the underlying parameter distribution.

White-box (model fine-tuning): Full access to model parameters. Techniques: full fine-tuning, LoRA, adapter layers, continued pre-training. Domain knowledge is incorporated directly into model weights. Most powerful but most resource-intensive — requires domain-specific datasets, compute, and expertise. Also carries the highest risk of Why do specialized models fail outside their domain?.

The access level is usually determined by organizational context rather than technical preference. API-only deployment (black-box) covers most enterprise use. Gradient access requires model weights (grey-box). Parameter modification requires infrastructure to train (white-box).

This taxonomy matters because practitioners often default to prompt-based approaches without recognizing that prompt optimization is bounded by Can prompt optimization teach models knowledge they lack?. When the required domain knowledge isn't in the model's training distribution, no amount of prompting will supply it — the tier must change.


Source: Domain Specialization

Related concepts in this collection

Concept map
15 direct connections · 156 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

domain specialization access taxonomy — black box grey box white box determines available techniques