ChatGPT is three years old today
A retrospective on ChatGPT's third anniversary, covering its surprising launch, initial internal skepticism, and unprecedented growth to 800 million users.
A retrospective on ChatGPT's third anniversary, covering its surprising launch, initial internal skepticism, and unprecedented growth to 800 million users.
Wikipedia's new guideline advises against using LLMs to generate new articles from scratch, highlighting limitations of AI in content creation.
Moonshot AI's Kimi K2 Thinking is a 1 trillion parameter open-weight model optimized for multi-step reasoning and long-running tool calls.
Explores the common practice of developers assigning personas to Large Language Models (LLMs) to better understand their quirks and behaviors.
A hands-on guide for JavaScript developers to learn Generative AI and LLMs through interactive lessons, projects, and a companion app.
An introduction to reasoning in Large Language Models, covering concepts like chain-of-thought and methods to improve LLM reasoning abilities.
Explores four main approaches to building and enhancing reasoning capabilities in Large Language Models (LLMs) for complex tasks.
Explains how multimodal LLMs work, compares recent models like Llama 3.2, and outlines two main architectural approaches for building them.
Explores methods for using and finetuning pretrained large language models, including feature-based approaches and parameter updates.
Exploring mixed-precision techniques to speed up large language model training and inference by up to 3x without losing accuracy.
A guide to parameter-efficient finetuning methods for large language models, covering techniques like prefix tuning and LLaMA-Adapters.
Guide to finetuning large language models on a single GPU using gradient accumulation to overcome memory limitations.
A curated reading list of key academic papers for understanding the development and architecture of large language models and transformers.
Analyzes the limitations of AI chatbots like ChatGPT in providing accurate technical answers and discusses the need for curated data and human experts.