LLM Notes

LLM Notes

A few notes and links for large language models (LLMs) as of Feb 2026.

Use Cases

Divergent Thinking

For divergent thinking, I use Anthropic’s Claude Opus 4.5 (via Claude Code).

I like its agency: it takes initiative and completes tasks without much back-and-forth unless you use Plan Mode.

Converging Thoughts → Writing Code

Once I have a plan, I switch to OpenAI’s GPT-5.2-Codex (via the Codex CLI).

Rough workflow:

  1. Claude Opus 4.5 for initiative and divergence (get something moving)
  2. GPT-5.2 Codex for implementation (write the code)

2026-Q1 Gen-AI Guide (Personal Workflow)

VendorModels (January 2026)Max ContextBest At
OpenAIGPT-5.2 (Codex) (system card (PDF))Code writing & implementation
 GPT-4.11M tokensCode & structured work, logic
 o3200kCreative writing, brainstorming, ideation
GoogleGemini 2.5 Pro1MCode, large codebases, large context prompts
 Gemini 2.5 Flash (docs)128kFast responses
AnthropicClaude 4 Sonnet200kCode & structured work, logic
 Claude Opus 4.5 (system card (PDF))Agentic workflows, ideation, divergence
 Claude 3.7 Sonnet200kCode & structured work, logic
MetaLlama 4 Scout (open)10M/1MOpen Weight LLM
xAIGrok 3130kReasoning, code & structured work, logic
 Grok 4256kReasoning, coding, multimodal
DeepSeekV3-0324, R1-052864kCost-sensitive experimentation

OpenAI

Anthropic

Previous Anthropic models - **Claude 4 Sonnet** - July 3, 2025 - **Claude 3.7 Sonnet** - February 24, 2025 ([system card](https://www.anthropic.com/claude-3-7-sonnet-system-card)) - **Claude 3.5 Sonnet** - June 20, 2024

Google

Previous Google models - **Gemini 2.5 Pro** - June 2025 ([model card](https://storage.googleapis.com/model-cards/documents/gemini-2.5-pro-preview.pdf))

Decoding OpenAI’s Labels

LabelTranslation
o*“Omni” = multimodal & agent-ready
4.1Latest GPT-4 weights
mini / nanoFewer parameters → lower cost & latency

Snapshot suffixes (e.g., -0425) indicate frozen weights date.

Notes

  1. Open-weight models (June 2025 - Feb 2026):
  2. synthetic.new - API provider for open-weight models (GLM-4.7, MiniMax M2.1).

  • “Context” refers to how many tokens the model can process in a single request. For Anthropic models, see their context windows documentation.

© Mark Norgren. Some rights reserved.

Build Date: 2026-02-21

976dcde