Best of Machine LearningOctober 2025

  1. 1
    Article
    Avatar of ecykgssvlzkqy1g7nxkjcFriedrich WT·28w

    AI Engineers then Vs Now

  2. 2
    Video
    Avatar of fireshipFireship·28w

    OpenAI’s new slop machine is open for business…

    OpenAI launched Sora 2, a video generation model that creates realistic videos with sound from text prompts. The platform functions as both a creation tool and social network with explore feeds, profiles, and invite-only access. The release follows Meta's similar Vibes feature, signaling a shift toward AI-generated content platforms. Sora 2 demonstrates significant improvements in physical accuracy and realism compared to previous video generation models, though it raises questions about the direction of AI development toward content creation rather than other applications.

  3. 3
    Article
    Avatar of ieeespectrumIEEE Spectrum·27w

    Wi-Fi Signal Tracks Heartbeat Without Wearables

    Researchers at UC Santa Cruz developed Pulse-Fi, a system that uses ambient Wi-Fi signals to monitor heart rate without wearables or cameras. The AI-powered approach runs on affordable devices like Raspberry Pi or ESP32 microcontrollers, filtering signal amplitude changes caused by heartbeats. Testing with over 100 participants showed less than 1.5 beats-per-minute error rate across various postures and distances up to 10 feet. The team is now working on multi-user support and exploring applications for sleep apnea and breathing rate monitoring.

  4. 4
    Article
    Avatar of hnHacker News·25w

    character-ai/Ovi

    Ovi is an open-source audio-video generation model that simultaneously creates synchronized 5-second videos and audio from text or text+image inputs. The 11B parameter model supports flexible resolutions (720×720 to 960×960), multiple aspect ratios, and includes a custom-trained 5B audio branch. It offers inference options for single or multi-GPU setups, includes memory optimization features like fp8 quantization and CPU offloading for 24GB GPUs, and provides integration with Gradio UI and ComfyUI. The model is based on research from Character AI and builds upon Wan2.2 for video and MMAudio for audio processing.

  5. 5
    Article
    Avatar of hnHacker News·25w

    apple/pico-banana-400k

    Apple released Pico-Banana-400K, a dataset containing approximately 400,000 text-image-edit triplets for training text-guided image editing models. The dataset includes 257K single-turn examples, 56K preference learning samples, and 72K multi-turn conversations, covering 35 edit operations across 8 semantic categories. Built using Gemini-2.5-Flash for instruction generation and the Nano-Banana model for editing, each edit undergoes automated quality evaluation. Source images come from Open Images, with edits spanning object manipulation, scene composition, stylistic changes, and photometric adjustments. The dataset is available under CC BY-NC-ND 4.0 license for non-commercial research use.

  6. 6
    Article
    Avatar of hnHacker News·28w

    The RAG Obituary: Killed by Agents, Buried by Context Windows

    RAG (Retrieval-Augmented Generation) architectures are becoming obsolete as LLM context windows expand dramatically from 4K to 2M+ tokens. The author argues that agentic search systems using simple tools like grep and filesystem navigation outperform complex RAG pipelines involving chunking, embeddings, hybrid search, and reranking. Drawing from experience building financial research platforms, they demonstrate how agents can navigate complete documents and follow cross-references naturally, eliminating the infrastructure burden and accuracy problems inherent in fragment-based retrieval. The shift from context scarcity to abundance fundamentally changes how AI systems should process information.

  7. 7
    Article
    Avatar of hnHacker News·24w

    Stop Citing AI

    Large language models like ChatGPT, Claude, and Gemini predict likely word sequences rather than provide factual information. These AI systems can generate convincing-sounding responses, but they lack source attribution and may produce inaccurate or unreliable information through hallucinations. Treating LLM outputs as authoritative sources is problematic, as they represent common word patterns rather than verified truths. The piece emphasizes the risks of over-trusting AI-generated content, particularly in critical domains like medicine and law.

  8. 8
    Article
    Avatar of hnHacker News·27w

    The AI bubble is 17 times the size of the dot-com frenzy - and four times subprime, this analyst argues

    MacroStrategy Partnership argues AI represents a bubble 17 times larger than the dot-com era and 4 times bigger than the 2008 housing crisis, based on Wicksellian economic theory measuring capital misallocation from artificially low interest rates. The analysis claims large language models have hit scaling limits, citing ChatGPT-5's $5 billion cost with minimal improvement over ChatGPT-4, low task completion rates at companies (1.5-34%), and declining AI adoption among large enterprises. The firm predicts this will trigger a deflationary recession similar to the early 1990s S&L crisis, recommending investors shift away from AI companies toward resources, emerging markets, gold, and short-dated Treasuries.

  9. 9
    Article
    Avatar of opensuseopenSUSE·27w

    GSoC 2025, Building a Semantic Search Engine for Any Video

    A GSoC 2025 project that built an end-to-end semantic video search engine capable of finding specific moments within videos using natural language queries. The system uses a two-part architecture: an ingestion pipeline that processes videos with AI models (TransNetV2, WhisperX, BLIP, VideoMAE) to extract shots, transcripts, captions, and actions, then segments them intelligently and enriches them with LLM-generated summaries; and a search application with FastAPI backend that performs hybrid text-visual searches using ChromaDB vector database and Reciprocal Rank Fusion for result ranking, paired with a Streamlit frontend for user interaction.

  10. 10
    Article
    Avatar of hnHacker News·27w

    Who needs git when you have 1M context windows?

    A developer accidentally lost code that improved their machine learning model by 5% after refactoring without committing changes. Unable to reproduce the results, they discovered that Gemini 2.5 Pro's 1M token context window had retained the original code from their development session, allowing them to recover the lost improvements through a simple prompt.

  11. 11
    Article
    Avatar of javarevisitedJavarevisited·24w

    I’ve Read 20+ Books on AI and LLM — Here Are My Top 5 Recommendations for 2026

    A curated list of five essential books for learning AI and LLM engineering, covering practical topics from building and fine-tuning models to production deployment. The recommendations include hands-on guides for prompt optimization, retrieval-augmented generation, model evaluation, infrastructure design, and understanding transformer architectures from scratch. Each book emphasizes production-ready engineering practices including monitoring, cost optimization, and system design rather than pure theory.

  12. 12
    Article
    Avatar of arstechnicaArs Technica·26w

    Nvidia sells tiny new computer that puts big AI on your desktop

    Nvidia launched the DGX Spark, a $4,000 desktop AI workstation featuring one petaflop of computing power and 128GB of unified memory in a compact form factor. The system can run AI models with up to 200 billion parameters locally and fine-tune models up to 70 billion parameters, addressing the need for developers who want to avoid cloud services. Built on the GB10 Grace Blackwell Superchip with ConnectX-7 200Gb/s networking, it targets AI developers working with large language models and media synthesis applications. Orders begin October 15 through Nvidia's website and select retail partners.

  13. 13
    Article
    Avatar of mlmMachine Learning Mastery·25w

    7 Must-Know Agentic AI Design Patterns

    Seven proven design patterns for building production-ready AI agents: ReAct (reasoning loops), Reflection (self-critique), Planning (task decomposition), Tool Use (external integrations), Multi-Agent Collaboration (specialized agents), Sequential Workflow (fixed pipelines), and Human-in-the-Loop (safety checkpoints). Each pattern addresses specific trade-offs between cost, latency, reliability, and complexity. The guide emphasizes starting simple with single agents and tool use, then evolving to more complex patterns only when clear limitations emerge. Includes practical decision framework based on workflow predictability, quality requirements, and task complexity.

  14. 14
    Article
    Avatar of huggingfaceHugging Face·24w

    huggingface_hub v1.0: Five Years of Building the Foundation of Open Machine Learning

    The huggingface_hub Python library has reached v1.0 after five years of development, now powering 200,000 dependent libraries and providing access to over 2 million models, 500,000 datasets, and 1 million Spaces. Major changes include migration from requests to httpx for modern HTTP infrastructure, a redesigned CLI replacing huggingface-cli with expanded features, and full adoption of hf_xet for file transfers with chunk-level deduplication. The release removes legacy patterns like the Git-based Repository class while maintaining backward compatibility for most ML libraries, though transformers v5 will be required for full v1.x support.

  15. 15
    Article
    Avatar of phProduct Hunt·26w

    nanochat: The best ChatGPT that $100 can buy

    nanochat is a minimal, full-stack LLM implementation by Andrej Karpathy in approximately 1000 lines of code. It enables running the complete pipeline—tokenization, pretraining, finetuning, evaluation, inference, and web UI—on a single 8XH100 node for under $1000. The project achieves competitive performance at the $100 tier model level while maintaining clean, hackable code designed to make LLM development accessible for learning purposes.

  16. 16
    Article
    Avatar of nvidiaNVIDIA·26w

    Elon Musk Gets Just-Launched NVIDIA DGX Spark: Petaflop AI Supercomputer Lands at SpaceX

    NVIDIA launched DGX Spark, a desktop-sized AI supercomputer delivering one petaflop of performance with 128GB unified memory, capable of running models up to 200 billion parameters locally. CEO Jensen Huang personally delivered the first unit to Elon Musk at SpaceX's Starbase facility. The system targets developers, researchers, and creators who need supercomputer-class AI performance in a portable form factor, with general availability starting October 15, 2025.

  17. 17
    Article
    Avatar of windsurfWindsurf·24w

    Introducing SWE-1.5: Our Fast Agent Model

    Windsurf releases SWE-1.5, a frontier-size AI model with hundreds of billions of parameters optimized for software engineering tasks. The model achieves near state-of-the-art coding performance while delivering inference speeds up to 950 tokens per second through a partnership with Cerebras—6x faster than Haiku 4.5 and 13x faster than Sonnet 4.5. Trained on real-world coding scenarios with input from senior engineers, SWE-1.5 focuses on writing clean, maintainable code and integrates tightly with Windsurf's agent experience. The model excels at exploring large codebases, full-stack development, and infrastructure work, reducing task completion times from 20+ seconds to under 5 seconds.

  18. 18
    Video
    Avatar of fireshipFireship·28w

    Alibaba is going all in on Qwen…

    Alibaba announced a $52 billion three-phase roadmap to artificial superintelligence at their Apsara conference, targeting completion by 2032. Key releases include Qwen 3 Max, a trillion-parameter model trained on 36 trillion tokens using mixture-of-experts architecture; Qwen 3VL, an open-source vision-language model that tops the Clockbench benchmark; and Qwen 3 Omni, a multimodal model capable of processing visual, audio, and text inputs. The roadmap progresses from generalized understanding through autonomous action to self-iteration with physical world integration.

  19. 19
    Article
    Avatar of gettothetopGet to the Top·24w

    AI Isn’t Replacing Writers. It’s Replacing the Internet.

    Oxford data reveals that over 50% of online text is now AI-generated, up from 5% in 2020, with projections suggesting 90% by next year. This shift raises concerns about model collapse—where AI systems trained on AI-generated content produce increasingly degraded output—and the "dead internet theory," where automated systems dominate online discourse. The trend prioritizes cost and speed over originality and truth, potentially creating a feedback loop of recycled, low-quality content that may ultimately drive demand for authentic human voices.

  20. 20
    Article
    Avatar of uxplanetUX Planet·24w

    UX 3.0

    UX 3.0 represents a paradigm shift from interface-centered design to intelligent ecosystem orchestration, where designers create experiences spanning interconnected devices and AI-powered systems. This evolution introduces four core pillars: ecosystem-based experiences across product lifecycles and platforms, human-AI symbiosis enabling predictive and contextual interactions, ethical considerations around transparency and fairness in AI systems, and co-creation methodologies that democratize the design process. Companies like Google, Netflix, and Spotify exemplify this approach by building adaptive systems that anticipate user needs, personalize experiences through machine learning, and maintain consistency across complex technological ecosystems while addressing challenges of algorithmic bias, privacy, and digital well-being.

  21. 21
    Article
    Avatar of mitMIT News·25w

    The student becomes the teacher

    MIT graduate student Titus Roesler overcame a challenging start from a rural background without AP classes to become an award-winning teaching assistant and mentor. His teaching work helped him master signal processing, where he now focuses on compressed sensing applications in high-frequency radio communications. Through roles as a TA for multiple classes and designing seminars, he developed expertise in source separation problems, including a project separating harmonies in Bach chorales using Python.

  22. 22
    Article
    Avatar of zerotomasteryZero To Mastery·26w

    AI wrote it. I cried.

  23. 23
    Article
    Avatar of dailydoseofdsDaily Dose of Data Science | Avi Chawla | Substack·25w

    ARQ: A New Structured Reasoning Approach for LLMs

    Researchers introduced Attentive Reasoning Queries (ARQs), a structured reasoning approach that prevents LLM hallucinations by guiding models through explicit, domain-specific questions encoded in JSON schemas. Unlike free-form techniques like Chain-of-Thought, ARQs force LLMs to follow controlled reasoning steps, achieving a 90.2% success rate compared to 86.1% for CoT. The approach is implemented in Parlant, an open-source framework for building instruction-following agents, where ARQs are integrated into guideline proposers, tool callers, and message generators to maintain alignment throughout multi-turn conversations.

  24. 24
    Article
    Avatar of hnHacker News·24w

    MoonshotAI/Kimi-Linear

    Kimi Linear introduces a hybrid linear attention architecture featuring Kimi Delta Attention (KDA), a refined version of Gated DeltaNet with improved gating mechanisms. The 48B parameter model (3B activated) supports 1M token context length, reduces KV cache requirements by 75%, and achieves 6x faster decoding throughput compared to traditional attention methods. Released as open-source with model checkpoints trained on 5.7T tokens, it demonstrates superior performance on long-context tasks while maintaining efficiency through a 3:1 KDA-to-global MLA ratio.

  25. 25
    Article
    Avatar of hnHacker News·25w

    Is Sora the Beginning of the End for OpenAI?

    OpenAI's release of Sora 2, a video generation model, includes a TikTok-style social app that creates AI-generated videos from text prompts. The app's focus on engagement-driven content and monetization through ads suggests a strategic shift from OpenAI's earlier positioning as a transformative AGI company. High operational costs and questionable content quality raise doubts about the app's viability. This pivot from revolutionary AI ambitions to consumer entertainment products may signal that OpenAI recognizes its technology won't deliver the immediate world-changing impact once promised.