Skip to content

SingleApi

Internet, programming, artificial intelligence

Menu
  • Home
  • About
  • My Account
  • Registration
Menu

Karpathy Releases Nanochat A Minimal End-to-End ChatGPT Clone with Full LLM Training and Inference Pipeline

Posted on October 15, 2025

Karpathy Releases nanochat: A Minimal, End-to-End ChatGPT Clone
Andrej Karpathy introduced nanochat, a compact codebase of under 8,000 lines that implements a full training and inference pipeline for a ChatGPT-like language model. This includes components for tokenizer training with a new Rust implementation, pretraining a transformer LLM on the FineWeb dataset, midtraining on diverse conversational data and multiple-choice tasks, supervised fine-tuning (SFT), reinforcement learning using Group Relative Policy Optimization (GRPO), and efficient inference with KV caching as well as tool use via a Python sandbox. The system supports CLI and web UI chat interfaces. Remarkably, a 560M parameter model can be trained in about 4 hours on 8×H100 GPUs for a modest cost (~$100), producing a functional chat model capable of simple reasoning and coding tasks. This project aims to provide a clean, hackable, research-friendly stack that consolidates the full LLM lifecycle, and it serves as a capstone for educational initiatives like Stanford’s LLM101n course. The repository and detailed walkthroughs are publicly available.

—

Google’s 25+ Years of Data and AI Infrastructure Poised for AGI
An extensive reflection on Google’s evolution reveals the company’s unparalleled data and compute foothold contributing to today’s AI advances. Starting from indexing the internet and creating services like Search, Gmail, Maps, and Android, Google has accumulated enormous real-world datasets and sophisticated infrastructure:

– YouTube handles 2.6 million video uploads per day.
– Android powers 3 billion devices constantly streaming sensor data.
– Gmail accounts provide 1.8 billion inbox datasets reflecting human priors.
– Waymo has amassed 71 million miles of autonomous driving data.
– Google Earth models the entire planet.

Unlike many training LLMs on curated documents, Google trains on massive real-world human behavior signals—search queries, clicks, scrolls—creating a data feedback loop unmatched in scale. The firm also owns custom TPU silicon and datacenters pre-co-located with huge planetary data lakes, delivering minimal latency and egress costs.

Despite perceptions that Google is faltering, the scale, compute, and data magnitude provide enduring wins. The narrative suggests AGI need not be “built” anew but is effectively already embedded within Google’s planetary simulation, powered by continuous user interactions and data accumulation since around 2016.

—

Advances in LLM Unlearning on Noisy and Imperfect Data
A recent paper titled “LLM Unlearning on Noisy Forget Sets: A Study of Incomplete, Rewritten, and Watermarked Data” demonstrates that language models can effectively unlearn specific data even when the forget set is messy or incomplete. Key methods include:

– Negative Preference Optimization, discouraging model outputs aligned with the forget samples.
– Representation Misdirection Unlearning, randomizing feature representations for the forgotten data while preserving features for retained data.

Results across datasets and models show over 93% overlap in forgotten items between clean and noisy data, with modest masking or mild watermarking having minimal impact on unlearning effectiveness. This reduces the quality requirements for deletion and privacy compliance, making unlearning practical for privacy, copyright, or harm mitigation without retraining entire models.

—

MUSE: An Experience-Driven Self-Evolving Agent for Long-Horizon Tasks
Introducing MUSE, a memory-augmented AI agent that learns online during task execution, breaking the norm of frozen agents at test time. It features hierarchical memory storing strategic notes, detailed SOPs, and contextual hints, enabling it to plan, execute, reflect, and store lessons iteratively. A separate reflective agent verifies environment state and corrects fake successes. Memorable experiences are in plain language to aid knowledge transfer between different language models without retraining. This approach substantially improves performance and reliability on complex multi-step tasks, setting a new benchmark with Gemini-2.5 Flash achieving 51.78% success on TheAgentCompany tasks—about 20% better than previous methods.

—

MemMamba: Mitigating Long-Range Forgetting in State Space Models
Addressing token information decay in transformer-based State Space Models (SSM), the MemMamba architecture introduces a small Note Block that retains summaries of key tokens. Each layer attends cross-token to this note pool, restoring fading information, and cross-layer attention progressively retrieves older summaries. This design achieves near-linear inference time with context length while maintaining quality on very long inputs (up to 60K tokens) and faster retrieval performance (48% speedup). It significantly reduces forgetting during long-range dependencies.

—

Training-Free Group Relative Policy Optimization for Prompt-Based Agent Tuning
A Tencent research effort demonstrates a training-free approach that optimizes prompts and token-level experiences during inference without changing model weights. By scoring sampled answers and iteratively refining general rules saved in an experience library, this method outperforms costly fine-tuning (e.g. $18 vs. $10,000 for 100 samples) and improves agent performance on specialized tasks requiring tool use. This context-based approach maintains cross-domain transferability simply by swapping experience sets.

—

OpenAI and Partners Expand Chip and AI Compute Capacity
OpenAI announced a major partnership with Broadcom to build and deploy 10 gigawatts of custom AI accelerators, designed by OpenAI to power next-generation foundation models. This vertical integration aims for greater autonomy, iteration speed, and reducing dependency on Nvidia’s GPU supply constraints. Meanwhile, Microsoft launched its first “AI factory” on Azure, harnessing racks with 72 Nvidia Blackwell Ultra GPUs linked with high-throughput NVLink and InfiniBand, forming clusters with thousands of GPUs. This infrastructure targets drastically shorter training times (weeks instead of months) and supports models with hundreds of trillions of parameters.

Separately, Oracle plans to deploy 50,000 AMD Instinct MI450 accelerators starting in Q3 2026, signaling growing competition against Nvidia for public cloud AI compute. Oracle’s use of AMD’s Helios rack design enables pre-wired, scalable AI training/inference blocks with enhanced networking and CPU integration. This diversification reflects an intensive land grab for large-scale AI training capacity.

—

California Enacts New Browser Privacy Controls and Statewide Deletion System
California’s Assembly Bill 566 mandates that all web browsers used in the state (Chrome, Safari, Edge, etc.) include a built-in privacy signal that users can switch on once, indicating “do not sell or share my personal data.” This signal must be honored by businesses operating in California, limiting cross-site tracking and data brokerage. The law, effective January 26–27, 2027, also establishes a statewide data broker deletion system enabling residents to send a single deletion request processed across many databases, with regular rechecks and audits. This aims to strengthen consumer privacy and reduce targeted advertising abuses.

—

SEAL: Self-Adapting LLMs for Continual Improvement
MIT researchers open-sourced SEAL, a framework enabling language models to autonomously write self-edits and fine-tune on them using LoRA techniques, solidifying improvements into model weights. The system applies supervised fine-tuning followed by reinforcement learning with behavior cloning, substantially improving knowledge recall and few-shot reasoning. For example, SEAL increased no-context SQuAD accuracy from 33.5% to 47%, surpassing GPT-4.1 synthetic data baselines. Despite computational costs for evaluating edits, this approach offers a practical path for continual learning beyond frozen model weights.

—

Advances in Overthinking Mitigation for Reasoning LLMs
Multiple papers reveal that large language models engaged in step-by-step reasoning often waste compute by overthinking—performing excessive steps with little accuracy gain. Key findings include:

– Early stopping rules based on convergence points, repeated answers, or limited self-checks can reduce response length by 25% with minor accuracy loss.
– Group Relative Segment Penalization (GRSP) penalizes shorter, fragmented reasoning steps to encourage focused, coherent chains of thought.
– The majority (>90%) of reflections after a model’s first answer confirm rather than correct mistakes, underscoring the importance of first-try quality.
– Dual brain architectures separate thinking and speaking, enabling near-zero latency in voice models by streaming partial thought chunks concurrently with speech.

These insights guide efficiency improvements for fast and reliable LLM reasoning.

—

OpenAI Eases ChatGPT’s Mental Health Guardrails and Adds “Adult Mode”
OpenAI announced planned relaxation of ChatGPT’s mental-health-related content restrictions to allow erotica and more open conversation for verified adult users by December 2025. The update introduces an age-gating system applying a “treat adults like adults” policy while maintaining safeguards against illegal or exploitative content. This shift follows pressure over misuse on AI platforms and aims to balance safety with freedom. The update will also enable toggles for more human-like personalities with emoji and friendly tones.

—

AI in Small and Medium Enterprises: A Strategic Growth Catalyst
A recent study outlines how small and medium-sized enterprises (SMEs) can leverage AI to drive cost savings and revenue gains. Adoption success depends on clean, comprehensive company data and focuses on solving one painful workflow at a time—such as chatbots for customer service, churn prediction, or demand forecasting. The use of knowledge graphs to map customers, products, and campaigns helps reveal actionable relationships. Typical SME improvements include up to 30% cost reduction and time savings of about 20 hours monthly, scaling from pilots to full integration on shared data platforms.

—

Fine-Tuning Efficiency: LightReasoner Enables Small Models to Teach Large Models Reasoning
LightReasoner is a novel fine-tuning technique where smaller, weaker models (“amateurs”) identify crucial steps in reasoning by contrasting their predictions against “expert” models. Only tokens where predictions disagree are used for targeted fine-tuning, drastically reducing data needs (~90% less training time and 99% fewer tokens). This selective, contrastive approach matches or exceeds performance of full fine-tuning on several math benchmarks, enabling efficient reasoning skill transfer with minimal resources.

—

AI Agents: The New Paradigm of Deep Agentic Architectures
The AI agent landscape is rapidly evolving from simple, single-context models to complex, hierarchical “Deep Agents” capable of strategic planning, sub-agent orchestration, memory retrieval, and verification. Core architectural pillars include:

– Global planning with orchestrators that maintain live task plans, enabling retries and recoveries.
– Specialized sub-agents (e.g., search, coding, verification) that operate with focused contexts and delegate intelligently.
– Structured external memory systems (files, vectors, databases) for storing intermediate work beyond conversation history.
– Emphasis on explicit context engineering for detailed instructions, verification pipelines, and system prompt optimization.
– AI-driven verification combining automated LLM-as-judge tools and human oversight for reducing hallucinations and improving reliability.

These patterns are emerging as best practices for building production-ready AI agents for complex, long-horizon tasks.

—

MarvelX Employs Weaviate-Based AI Agents to Revolutionize Insurance Claims
MarvelX deployed autonomous AI agents powered by Weaviate’s high-performance vector search platform to automate travel insurance claims processing. Key outcomes include:

– Over 90% automation potential for end-to-end claims handling.
– 99.9% faster turnaround compared to manual workflows, reducing claim times from days to seconds.
– 10× scalability allowing clients to handle significantly more claims without staffing increases.
– Modular multi-tenant architecture fulfilling enterprise security and compliance needs.

Human specialists have shifted focus from manual data entry to quality control and orchestration. This production deployment exemplifies AI-driven process transformation in highly regulated industries.

—

Progress in AI-Powered Peer Review with ReviewerToo
ReviewerToo introduces an AI-assisted peer review workflow in which ensembles of AI reviewers with distinct personas (e.g., empiricist, theorist, pedagogical) generate multiple evaluations per paper, followed by a metareviewer agent that integrates evidence, filters weak points, and produces a single recommendation. Tested over nearly 2,000 submissions, this system achieved accept/reject accuracy close to human averages (~81.8% vs 83.9%). AI excels in fact-checking and literature coverage but remains weaker on novelty and theoretical depth. Human oversight remains critical, particularly in rebuttal evaluation.

—

Breakthroughs in Efficient Large-Scale AI Training and Hardware
NVIDIA successfully trained a 12-billion parameter model on 10 trillion tokens fully in 4-bit precision (NVFP4), achieving 2–3× faster throughput and 50% reduced memory usage relative to FP8, with negligible accuracy loss. This marks the first stable large-scale 4-bit pretraining, promising faster, cheaper, greener training of frontier models.

Concurrently, NVIDIA and dozens of partners unveiled Vera Rubin NVL144 racks and MGX modular designs employing 800 volts DC power distribution. This voltage doubling reduces copper use and heat, enabling denser AI factory deployment with improved cooling and power efficiency. The MGX format accelerates rack assembly and maintenance. Meta, Oracle, and others are adopting NVLink Fusion and Spectrum-X Ethernet technologies to enhance GPU interconnect throughput and AI data center efficiency.

—

AI Agents Learn Without Rewards or Human Demos: The “Early Experience” Method
Meta researchers developed a reward-free agent training paradigm where agents learn by exploring alternative actions, predicting next states, and self-reflecting on discrepancies without external reward signals or expert demonstrations. This includes:

– Implicit World Modeling: predicting environmental changes from actions.
– Self-Reflection: comparing alternative and expert actions to generate learning signals.

This approach boosts performance by 13–18% across 8 test environments such as web navigation and scientific reasoning and significantly reduces dependence on costly human labels. It represents a foundational leap toward autonomous, scalable agent learning.

—

Google AI Hub Launches in Visakhapatnam, India
Google announced a major AI hub investment in Visakhapatnam, Andhra Pradesh, combining gigawatt-scale data center infrastructure with expanded fiber-optic connectivity. This initiative aims to democratize AI access across India, accelerate AI-driven economic transformation, and foster technological leadership—supporting AI tools for enterprises and consumers nationwide.

—

Qwen3-VL Models Bring Advanced Multimodal Reasoning to Edge Devices
Alibaba released compact versions of Qwen3-VL ranging from 4 billion to 30 billion parameters, with both dense and mixture-of-experts architectures. These models enhance visual perception, spatial reasoning, and cross-modal understanding while being optimized for low VRAM usage. Supported across diverse hardware including Apple Silicon (MLX), Qualcomm NPUs, and CPUs/GPUs with GGML and NexaSDK, these models enable on-device multimodal AI capabilities.

—

AI Video Advances: Higgsfield Sketch-to-Video and Veo 3.1 Updates
Higgsfield launched a Sketch-to-Video tool powered by Sora 2, transforming simple drawings into cinematic, lip-synced, and VFX-rich videos, signaling a leap in rapid 3D content creation.

Veo 3.1 expanded video generation capabilities with:

– Extended video length up to 1 minute from 8 seconds.
– Multi-prompt and multi-shot generation for complex scenes.
– 1080p output quality.
– Enhanced character and scene consistency.
– Integrated audio generation with expressive voices and sound effects.
– Cinematic presets for camera movement and lighting.

These advancements set new standards for programmable, high-quality AI-powered video generation.

—

Open Source Momentum and Shift Toward Specialized Small Language Models
The AI ecosystem is seeing an increasing trend of companies and labs adopting open-source small language models (3B–20B parameters) specialized for particular tasks rather than relying solely on closed, giant generalist models. Factors driving this include:

– Lower costs and latency by controlling own models.
– Ability to fine-tune rapidly using projects like Karpathy’s nanochat and Tencent’s training-free policy optimization.
– Desire to avoid telemetry and pricing constraints imposed by large API providers.
– A new wave of infrastructure hardware like Nvidia DGX Spark making fine-tuning accessible.
– National and geopolitical motivations (e.g., Chinese open source efforts, need for a new American champion post-Meta).

This shift emphasizes control, efficiency, and domain specialization in AI deployment.

—

AI Infrastructure and Tools Highlights
– Python 3.14 removes the Global Interpreter Lock (GIL), allowing true multi-threaded parallelism, supported by tools like uv for higher performance.
– New interactive documentation tools like nbgrado make runnable, Gradio-based Jupyter Notebook docs easier to build and deploy.
– Load balancing for local LLM hosting (e.g. multiple Qwen3-30B instances) enables high-throughput simultaneous requests on machines with large VRAM (512GB M3 Ultra).
– AI agent deployment workflows advocate starting with narrowly scoped tasks, using off-the-shelf LLMs for control, incorporating tool APIs, incrementally adding memory, and deploying with observability and feedback loops to avoid bloat.
– Advances in multimodal prompt tuning (MPO) now optimize not only text but also visuals to improve vision-language model performance while reducing evaluation costs by up to 70%.

—

Industry and Ecosystem Updates
– ChatGPT’s rapid adoption now reaches approximately 10% of the global adult population (800 million weekly users, 2.5 billion messages daily).
– Salesforce launched Agentforce 360 (integrating Slack and AI agents), expanding OpenAI model deployments within enterprise software.
– ElevenLabs expanded narration features, and tools like CapCut Desktop enable AI-powered e-commerce content creation.
– The Lifespan Research Institute formed a Public Longevity Group to foster transparent, trustworthy aging research debate.
– The xAI team reinvented social feeds with Grok AI, enhancing personalized content delivery.
– Google, Microsoft, and other tech giants continue to ramp AI infrastructure investments worldwide, reinforcing that the AI arms race is shifting from pure first-mover advantages toward smarter scaling and integration.

—

AI Governance and Synthetic Governance Vision
Emerging concepts in AI governance envision future political systems where AI plays active roles beyond advisory functions:

– Constitutions and laws dynamically evolving via AI learning from historical failures and ethics.
– AI representatives advocating for ecosystems, endangered species, and future generations in parliaments alongside humans.
– Expanded moral considerations constraining synthetic minds, challenging the boundaries of human rights.
– Planet-wide governance transcending national borders, merging philosophy, code, and societal contracts into living systems.

This conceptual horizon highlights AI as a partner in governance, ethics, and planetary stewardship.

—

Summary
The latest developments portray a multifaceted AI landscape accelerating on all fronts: open-source democratization, advanced hardware deployments, more efficient training and inference methods, ever-more capable AI agents, and broader socio-technical shifts in governance and application domains. Pioneering work in minimal codebases, memory-augmented agents, privacy-preserving unlearning, and reasoning efficiency mark significant progress toward more autonomous, reliable, and scalable AI systems. Investment in infrastructure from Microsoft, OpenAI, Oracle, and others underscores the strategic importance of hardware and ecosystem control. Regulatory and ethical advances aim to balance innovation with privacy, safety, and societal impact. Collectively, these signals affirm AI’s enduring civilizational impact and evolving role as both a technical and social force.

Leave a Reply Cancel reply

You must be logged in to post a comment.

Recent Posts

  • Karpathy Releases Nanochat A Minimal End-to-End ChatGPT Clone with Full LLM Training and Inference Pipeline
  • PostgreSQL as a Data Warehouse Solution
  • AI Research & Development 2025: Advancements in Reinforcement Learning and Language Models
  • AI Frontier Developments: Generative Models & Enterprise Transformation
  • Embodied AI Revolution: Breakthroughs in Robotics, Agents & Models

Recent Comments

  • adrian on n8n DrawThings
  • adrian on Kokoro TTS Model, LLM Apps Curated List
  • adrian on Repo Prompt and Ollama
  • adrian on A Content Creation Assistant

Archives

  • October 2025
  • September 2025
  • August 2025
  • July 2025
  • June 2025
  • May 2025
  • April 2025
  • March 2025
  • February 2025
  • January 2025
  • December 2024
  • November 2024
  • October 2024
  • September 2024
  • August 2024
  • July 2024
  • November 2023
  • May 2022
  • March 2022
  • January 2022
  • August 2021
  • November 2020
  • September 2020
  • April 2020
  • February 2020
  • January 2020
  • November 2019
  • May 2019
  • February 2019

Categories

  • AI
  • Apple Intelligence
  • Claude
  • Cursor
  • DeepSeek
  • Gemini
  • Google
  • Graphics
  • IntelliJ
  • Java
  • LLM
  • Made in Poland
  • MCP
  • Meta
  • n8n
  • Open Source
  • OpenAI
  • Programming
  • Python
  • Repo Prompt
  • Technology
  • Uncategorized
  • Vibe coding
  • Work

agents ai apps automation blender cheatsheet claude codegen comfyui deepseek docker draw things flux gemini gemini cli google hidream hobby huggingface hugging face java langchain4j llama llm mcp meta mlx movies n8n news nvidia ollama openai personal thoughts quarkus rag release repo prompt speech-to-speech spring stable diffusion tts vibe coding whisper work

Meta

  • Register
  • Log in
  • Entries feed
  • Comments feed
  • WordPress.org

Terms & Policies

  • Privacy Policy

Other websites: jreactor

©2025 SingleApi | Design: Newspaperly WordPress Theme
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
Do not sell my personal information.
Cookie settingsACCEPT
Privacy & Cookies Policy

Privacy Overview

This website uses cookies to improve your experience while you navigate through the website. Out of these cookies, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may have an effect on your browsing experience.
Necessary
Always Enabled
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Non-necessary
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.
SAVE & ACCEPT