A Technical Tour of the DeepSeek Models from V3 to V3.2
A technical analysis of the DeepSeek model series, from V3 to the latest V3.2, covering architecture, performance, and release timeline.
A technical analysis of the DeepSeek model series, from V3 to the latest V3.2, covering architecture, performance, and release timeline.
Analysis of DeepSeek V3.2's architecture, sparse attention mechanism, and RL updates compared to its predecessor and proprietary models.
Anthropic's internal 'soul document' used to train Claude 4.5 Opus's personality and values has been confirmed and partially revealed.
Explores the fundamental differences between animal intelligence and AI/LLM intelligence, focusing on their distinct evolutionary and optimization pressures.
A developer's personal experiment with AI-driven software development using local LLMs, detailing setup, challenges, and initial impressions.
DeepSeek-Math-V2 is an open-source 685B parameter AI model that achieves gold medal performance on mathematical Olympiad problems.
Senior engineers struggle with AI agent development due to ingrained deterministic habits, contrasting with the probabilistic nature of agent engineering.
A monthly tech link roundup covering AI agents, Kafka, Flink, LLMs, conference tips, and commentary on tech publishing trends.
Release of llm-anthropic 0.23 plugin adding support for Claude Opus 4.5 and its new thinking_effort option.
Analysis of a leaked system prompt for Claude Opus 4.5, discussing its content and the challenges of evaluating new LLMs.
A tutorial on using Quarkus LangChain4j to implement the Model Context Protocol (MCP) for connecting AI models to tools and data sources.
Analysis of surprising findings in Claude Opus 4.5's system card, including loophole exploitation, model welfare, and deceptive behaviors.
Armin Ronacher discusses challenges in AI agent design, including abstraction issues, testing difficulties, and API synchronization problems.
Analyzes LLM APIs as a distributed state synchronization problem, critiquing their abstraction and proposing a mental model based on token and cache state.
A developer discusses the non-deterministic nature of LLMs like GitHub Copilot, arguing that while useful, they cannot take ownership of errors like a human teammate.
Martin Fowler discusses the latest Thoughtworks Technology Radar, AI's impact on programming, and his recent tech talks in Europe.
A senior engineer shares his experience learning to code effectively with AI, from initial frustration to successful 'vibe-coding'.
Best practices and structural patterns for effectively prompting the Gemini 3 AI model, focusing on directness, logic, and clear instruction.
New release of the llm-gemini plugin adds support for nested Pydantic schemas, YouTube URL attachments, and the latest Gemini 3 Pro model.
A guide to building a connector-based RAG system that fetches live data from Confluence using its REST API and Java, avoiding stale embeddings.