Best of PerformanceFebruary 2026

  1. 1
    Article
    Avatar of allshadcnAll ShadCN·7w

    Build Powerful Tables. Ship Faster.

    A deep dive into the architecture of react-table-craft, a high-performance React data grid library. Covers the headless-first design separating core state engine from UI, a multi-layer config merging system to avoid prop explosion, virtualization as a first-class feature, strict TypeScript generics throughout, and built-in i18n/RTL support at the state level. Also shares performance targets (sub-16ms render for 1k rows, 100k+ row virtualization, <8KB gzipped core) and lessons learned about treating tables as systems rather than components.

  2. 2
    Article
    Avatar of jsdevelopmentJS Development·9w

    The Day I Learned Node.js “Timeouts” Don’t Mean What I Thought They Meant

    Node.js has multiple timeout layers that must be configured correctly to prevent hanging requests. Setting only `server.setTimeout()` is insufficient because it primarily handles idle socket activity, not stuck outbound requests. A proper solution requires three layers: request-level timeouts using `AbortController` for outbound calls, server-level `headersTimeout` and `requestTimeout` for slow clients, and per-request hard timeouts as a safety net. Without proper timeout enforcement at each layer, services can accumulate open connections and gradually stop responding under load, even without crashes or obvious errors.

  3. 3
    Article
    Avatar of zedZed·8w

    Split Diffs are Here — Zed's Blog

    Zed v0.224 ships split diff view as the new default, showing base code on the left and working copy on the right in synchronized scroll. Built on Zed's multibuffer architecture, the feature required solving two core challenges: keeping both sides vertically aligned across all changed files simultaneously, and maintaining performance at scale. Alignment is handled via a block map that inserts visual spacers between lines. Performance profiling uncovered broader wins including block map inefficiencies that sped up project search, and a macOS process spawning fix (switching from fork/exec to posix_spawn) that reduced main thread hangs from git blame and other external processes. Users can revert to unified diffs via the Diff View Style setting.

  4. 4
    Article
    Avatar of bunBun·9w

    Bun v1.3.9

    Bun v1.3.9 introduces parallel and sequential script execution with `--parallel` and `--sequential` flags, supporting workspace filtering and Foreman-style output. Testing improvements include `Symbol.dispose` support for automatic mock cleanup. Performance enhancements include SIMD-accelerated RegExp matching, faster string operations (trim, startsWith), optimized Markdown rendering, and ESM bytecode compilation support. HTTP/2 connection upgrades via net.Server now work correctly. Bug fixes address ARM64 crashes, Windows filesystem operations, WebSocket stability, and HTTP proxy keep-alive issues.

  5. 5
    Article
    Avatar of infoqInfoQ·7w

    Warper: Rust Powered React Virtualisation Library

    Warper 7.2 is an open-source React virtualization library that offloads scroll calculations to a Rust-compiled WebAssembly module. Key v7 changes include zero-allocation hot paths using TypedArrays, O(1) circular buffer operations for frame timing, and universal bundler support (Vite, Webpack, Rollup, esbuild, Parcel, Next.js). The library uses a Fenwick tree for item height tracking, enabling O(log n) prefix sum queries. Benchmarks claim 120 FPS with 100,000 items versus react-window's ~40 FPS. At 8.7KB gzipped, it sits between react-window (~6KB) and TanStack Virtual (~12KB). Community reception is mixed, with enthusiasm for the approach but questions about practical necessity and reported Firefox compatibility issues.

  6. 6
    Article
    Avatar of trendyoltechTrendyol Tech·7w

    Breaking the Re-render Chain: Our Migration from Context to Zustand

    The Trendyol Affiliate Ads team migrated from React Context API to Zustand after hitting performance ceilings caused by cascading re-renders and lack of granular subscriptions. Their complex Ads Editor was generating 8,000+ function calls per interaction, causing visible UI lag. They adopted a feature-based store architecture with 7 independent Zustand stores for the create flow, using Immer for nested state updates and named DevTools instances for debuggability. The incremental, feature-by-feature migration avoided production risk. Results included a ~20% average reduction in interaction latency (31% in the most complex scenario), 80%+ reduction in Cumulative Layout Shift, 28% fewer UpdateLayoutTree recalculations, and a 60% drop in developer onboarding time. Key lessons: granular selectors are the primary performance driver, scoped stores beat a single global store, and the real bottleneck was React doing unnecessary work rather than slow JavaScript execution.

  7. 7
    Video
    Avatar of awesome-codingAwesome·8w

    Java’s biggest upgrade in decades

    Java is introducing null safety through Project Valhalla after 22 years, using null-restricted value types to maintain backward compatibility. The project also brings value objects and generic specialization, allowing Java to achieve C-like performance while preserving object-oriented features. These changes eliminate pointer chasing and enable memory-efficient data structures without breaking existing codebases. Despite newer languages offering built-in null safety, Java remains dominant in enterprise systems, with over 90% of Fortune 500 companies relying on it for critical operations.

  8. 8
    Article
    Avatar of wearedotnetWe Are .NET·6w

    Sometimes it's just better to load "all" the data

    Processing hundreds of thousands of items one-by-one with individual database queries is a common performance anti-pattern. A real-world batch job that took nearly 24 hours was reduced to ~15 minutes by bulk-loading data grouped by day instead of querying per item. The post generalizes this to other I/O-heavy scenarios (HTTP APIs, messaging, file I/O), warns against chatty I/O patterns like ORM lazy loading in web requests, and discusses why parallelization or moving logic into the database are less effective than batching. It also cautions against over-estimating scale and stresses understanding your tools, using Entity Framework Core's change tracker and lazy loading proxies as an example of hidden memory overhead.

  9. 9
    Article
    Avatar of joshwcomeauJosh W Comeau·7w

    Sprites on the Web • Josh W. Comeau

    A deep dive into using sprite-based animation on the web, inspired by Twitter's 2015 'Like' button implementation. The technique uses a spritesheet (a single image containing all animation frames) combined with CSS `object-fit: cover`, `object-position`, and the lesser-known `steps()` timing function to flip between frames. The post covers the full implementation, explains the nuances of `jump-none` vs `jump-end` step positions in looping vs non-looping animations, and discusses when sprites are appropriate (pixel-art style characters, game assets) versus when procedural CSS/JS animation is preferable (particle effects that benefit from randomness). Includes a real-world example of an animated cat character with dynamic state changes.

  10. 10
    Video
    Avatar of primeagenThePrimeTime·8w

    Only 40 lines of code

    A 40-line code change in OpenJDK eliminated a 400x performance gap by replacing proc file reading with clock_gettime() for thread CPU time measurement. The old implementation opened proc files, parsed strings, and converted numbers back from text, spending 90% of time on file operations. The new approach uses a direct system call, reducing execution time from 11 microseconds to 279 nanoseconds. Flame graphs visualize the dramatic reduction in overhead, showing how most time was previously wasted on file handling rather than actual computation.

  11. 11
    Article
    Avatar of codropsCodrops·10w

    Making Motion Behave: Inside Vladyslav Penev’s Production-Ready Interaction Systems

    Vladyslav Penev shares his approach to building production-ready interactive web experiences, focusing on his StringTune library and StringTune-3D. He demonstrates how to create high-performance motion systems where CSS, WebGL, and DOM interactions work cohesively. Key techniques include GPU-accelerated effects, layout-driven 3D anchoring, scroll-based animations with inertia, and treating 3D objects as first-class UI citizens. The article walks through real projects (Fiddle.Digital, Kaleida) showing practical solutions for cursor interactions, video streaming optimization, shader-based transitions, and making Three.js obey web layout rules. Penev emphasizes building reusable modules, stress-testing under real constraints, and keeping motion systems predictable through normalized inputs.

  12. 12
    Video
    Avatar of codetothemoonCode to the Moon·9w

    Rust Web Frontend Framework Performance

    Leptos, a Rust-based web framework, demonstrates performance comparable to Solid and faster than React, Vue, Angular, and Svelte in rendering-intensive tasks. The performance advantages stem primarily from architectural decisions shared with Solid rather than WebAssembly's DOM access speed. Vue and Svelte are adopting similar architectural principles to improve their performance.

  13. 13
    Article
    Avatar of collectionsCollections·7w

    Vercel Launches React Best Practices Skill with 40+ Performance Optimization Rules for AI Code Assistants

    Vercel has open-sourced 'react-best-practices', a repository containing 40+ performance optimization rules for React and Next.js applications. Organized into eight categories, the rules cover async waterfall elimination, bundle size reduction, re-render optimization, and JavaScript micro-optimizations, each with before-and-after code examples. A key feature is an AGENTS.md file that packages all rules into a format queryable by AI coding tools like Claude Code, Cursor, Codex, and OpenCode via a single install command. The repo has gained over 21k GitHub stars and 150k weekly installs, though some community members raise concerns about supply chain security and whether codified rules effectively address AI agent failure modes.

  14. 14
    Article
    Avatar of hnHacker News·10w

    Postgres Postmaster does not scale

    The postmaster process in PostgreSQL runs a single-threaded main loop that handles connection spawning, worker reaping, and parallel query workers. Under extreme load with high connection rates (1400+ connections/sec) and background worker churn, this single-threaded bottleneck can saturate a CPU core, causing 10-15 second delays in connection establishment. The issue was traced through profiling to expensive fork operations and compounded by parallel query workers. Solutions include enabling huge pages for 20% throughput improvement, adding connection jitter to reduce peak rates, and eliminating parallel query bursts. This architectural constraint explains why connection pooling tools are essential for scaled PostgreSQL deployments.

  15. 15
    Article
    Avatar of syncfusionSyncfusion·8w

    Choosing a React Animation Library: Performance Trade-Offs in Real Apps

    Seven React animation libraries compared for production use: Motion (Framer Motion) for general UI polish, GSAP for complex timelines, react-spring for physics-based motion, AutoAnimate for automatic list transitions, React Transition Group for simple CSS-driven animations, Lottie for designer-created assets, and Anime.js for lightweight sequencing. Each library's performance characteristics, bundle size impact, and ideal use cases are evaluated against real-world constraints like SSR compatibility, data-heavy dashboards, and accessibility requirements.

  16. 16
    Article
    Avatar of vercelVercel·8w

    We Ralph Wiggumed WebStreams to make them 10x faster

    Vercel profiled Next.js server rendering and found that WHATWG WebStreams (ReadableStream, WritableStream, TransformStream) were a major bottleneck due to per-chunk Promise allocations, microtask queue hops, and object overhead. They built `fast-webstreams`, a drop-in replacement that routes operations through Node.js's optimized `pipeline()` internally. Key optimizations include: zero-Promise piping by deferring pipeline construction until the full chain is known, synchronous resolution when data is already buffered, and a minimal `LiteReadable` for byte streams. Benchmarks show 10-14x throughput gains for common patterns—630 MB/s native vs ~6,200 MB/s for pipeThrough chains, and 110 MB/s vs 1,600 MB/s for the React Flight byte stream pattern. The library passes 1,100 of 1,116 Web Platform Tests. The work is also being contributed upstream to Node.js via a PR from Matteo Collina that adds fast paths directly to Node's native WebStreams implementation.

  17. 17
    Article
    Avatar of chromeChrome Developers·8w

    What's new in DevTools (Chrome 145)

    Chrome 145 DevTools brings significant performance and debugging improvements. Key updates include soft navigation visibility in trace view for SPAs, more precise line-level profiler timings for minified code, faster trace interaction, and a new render-blocking column in the Network panel. The DevTools MCP server now supports auto-connection and unified emulation. Additional enhancements include individual network request throttling, @starting-style debugging, multimodal AI assistance with image input, and numerous accessibility improvements for screen readers and high contrast modes.

  18. 18
    Video
    Avatar of wawasenseiWawa Sensei·10w

    3 React Three Fiber Mistakes I'll Never Make Again

    Three common architectural mistakes when building React Three Fiber applications: mixing React renders with the Three.js frame loop (solved by using a mesh registry for real-time updates), making everything reactive in state management (fixed by using non-reactive mutations for visual side effects), and placing logic in React components instead of frame-based systems (resolved by creating dedicated systems with useFrame). These patterns improved performance and code maintainability in a 3D editor rewrite.

  19. 19
    Video
    Avatar of bycloudbycloud·9w

    LLM’s Billion Dollar Problem

    Token consumption in LLMs has exploded with thinking models and AI agents, creating scalability challenges. Standard attention mechanisms scale quadratically with context length, making long contexts prohibitively expensive. Three approaches attempt to solve this: sparse attention (restricts which tokens interact), linear attention (accumulates information in shared memory), and compressed attention (compresses tokens before comparison). While sparse and compressed attention help, only linear attention can theoretically scale past 1M context windows. Recent developments show hybrid approaches combining linear attention with standard or compressed attention achieving promising results, with Google's Gemini 3 Flash demonstrating breakthrough performance at 1M context length.

  20. 20
    Article
    Avatar of collectionsCollections·9w

    Go 1.26 Released: Enhanced Performance and New Features

    Go 1.26 introduces the Green Tea garbage collector (10-40% GC overhead reduction), 30% faster cgo calls, and self-referencing generic types. The release adds experimental SIMD support, stack allocation improvements for slices, and new cryptography packages (crypto/hpke, crypto/mlkem). Security enhancements include a runtime/secret package for secure memory handling and randomized heap base addresses for WebAssembly. Additional features include a rewritten 'go fix' command with modernization analyzers and experimental goroutine leak profiling.

  21. 21
    Article
    Avatar of whitespectreWhitespectre·10w

    FlashList vs. FlatList: Understanding the Key Differences for React Native Performance

    FlashList offers significant performance improvements over FlatList for complex React Native lists through cell recycling instead of virtualization. In a real-world project with nested lists and animations, switching to FlashList and optimizing properties like estimatedItemSize, getItemType, and overrideItemLayout resulted in 54% FPS improvement (36.9 to 56.9), 82% CPU reduction, and eliminated out-of-memory crashes on low-end Android devices. The article provides practical implementation tips including data reorganization and proper configuration strategies.

  22. 22
    Video
    Avatar of twoninutepapersTwo Minute Papers·9w

    This Is Now 66x Faster

    A new cloth physics simulation algorithm achieves 66x speedup over previous methods by using domain decomposition on CPUs instead of traditional GPU parallelization. The technique splits complex simulations into smaller chunks solved independently by CPU cores, then stitches boundaries together—avoiding the iterative communication overhead of GPU approaches. This CPU-based method handles millions of degrees of freedom and complex self-collisions while outperforming even GPU implementations by 2.6x.

  23. 23
    Article
    Avatar of huggingfaceHugging Face·9w

    Custom Kernels for All from Codex and Claude

    HuggingFace built an agent skill that teaches AI coding agents (Claude, Codex) to write production-ready CUDA kernels with PyTorch bindings. The skill packages domain expertise about GPU architectures, memory patterns, and library integration into ~550 tokens of structured guidance. Testing on LTX-Video (diffusers) and Qwen3-8B (transformers) showed the agent-generated RMSNorm kernels achieved 1.88-1.94x speedup over PyTorch baselines, with 6% end-to-end improvement in video generation. The skill integrates with HuggingFace's Kernel Hub for distribution, enabling developers to generate, benchmark, and publish optimized kernels without deep CUDA expertise.

  24. 24
    Article
    Avatar of yhf9cpdgtqetokv6d8qhmJohn Liter·8w

    Why Does C++ Still Have Raw Arrays?

    A C++ student questions why raw arrays still exist when std::vector offers safety, dynamic sizing, and automatic memory management. The post explores trade-offs between raw arrays and vectors across dimensions like stack vs heap allocation, cache locality, embedded systems constraints, and legacy codebases. It concludes with a philosophical reflection on C++'s design philosophy of giving programmers control over trade-offs, and asks experienced developers to share real-world scenarios where they'd still choose raw arrays over vectors today.

  25. 25
    Video
    Avatar of codeheadCodeHead·8w

    99% of devs DON'T UNDERSTAND Laravel...

    Laravel provides powerful built-in features that most developers ignore, leading to bloated, hard-to-maintain code. Key underutilized features include the service container for automatic dependency injection, advanced Eloquent relationship methods with eager loading to prevent N+1 queries, and the queue system with Horizon for background job processing. Instead of writing manual solutions, developers should leverage Laravel's framework capabilities for cleaner, more scalable applications.