Aggregate & Group: Verified Benchmarks. 10 Million Rows in 34 Seconds.
Full advanced mode with 19 functions (MEDIAN, P90, STDDEV, COUNT_DISTINCT) on a 1.2GB CSV. Simple mode hits 26 seconds. All in your browser — never uploaded.
Throughput by File Size — Simple vs Advanced Mode
| File Size | Simple Mode (5 functions) | Advanced Mode (19 functions) | Notes |
|---|---|---|---|
| 1K rows | ~25K rows/sec | ~18K rows/sec | Startup overhead dominates at small sizes (worker init, file read) |
| 10K rows | ~185K rows/sec | ~130K rows/sec | Worker pipeline warming up, chunk batching beginning to amortize |
| 100K rows | ~310K rows/sec | ~210K rows/sec | Typical CRM or financial export batch |
| 1M rows | ~390K rows/sec | ~265K rows/sec | 5 GROUP BY dimensions, 3 metric columns |
| 5M rows | ~400K rows/sec | ~280K rows/sec | Streaming chunks fully pipelined, peak throughput approaching |
| 10M rowsVerified | ~380K rows/sec (~26s) | ~294K rows/sec (~34s) | Verified benchmark — peak throughput, memory-bounded at scale |
| ~1.2GB file | ~365K rows/sec | ~270K rows/sec | Near browser-memory ceiling — results vary significantly by available RAM |
Advanced mode: All 19 functions including MEDIAN, P25–P95, MODE, STDDEV, VARIANCE, FIRST, LAST, COUNT_DISTINCT, CONCAT_UNIQUE — requires value-storage arrays per group for accurate computation.
All benchmarks: Chrome 132, Windows 11, Intel i7-12700K, 32GB RAM, February 2026. Results vary ±15–20%.
Rows Per Second — Visual Comparison
Simple Mode — 5 Functions (COUNT, SUM, AVG, MIN, MAX)
Higher values indicate better performance (faster processing)
Advanced Mode — 19 Functions (includes MEDIAN, Percentiles, STDDEV)
Higher values indicate better performance (faster processing)
Performance improves with larger files due to 10MB chunk batching amortizing startup overhead. Results vary ±15–20% by hardware, browser, and function selection.
Benchmark Methodology & Test Configuration
Why Browser-Based Aggregation Outperforms Excel on Large Files
What Excel Pivot Tables Do
What SplitForge Does
The architectural result: Excel's performance ceiling is determined by worksheet materialization cost — rendering 1M+ rows into a grid exhausts memory before aggregation even begins. SplitForge's ceiling is determined by HashMap cardinality (number of unique group combinations) and function selection. Processing 10M rows with 500 unique groups requires storing ~500 aggregate records — not 10 million rows. That's why 10M rows processes in 26–34 seconds while Excel cannot open the file at all.
Time Savings Calculator
Baseline: Excel Pivot on large file ≈ 20 min per analysis. SplitForge ≈ 2 min.
20-minute baseline derived from internal timing of Excel Pivot Table setup and processing on 500K–1M row exports (field collapse, recalculate, filter copy). Results vary by file size and analyst experience.
Function Performance Characteristics
Online algorithms — maintain running state without storing values. No additional RAM per group beyond the running total. Performance is essentially constant regardless of group size.
STDDEV/VARIANCE use Welford's algorithm — O(1) memory but O(n) computation. COUNT_DISTINCT and CONCAT_UNIQUE maintain a Set per group. FIRST/LAST require tracking insertion order.
Require storing all values per group to compute the true median or percentile. At 10M rows with 500 groups, this means potentially millions of stored values. Sorted at finalization — O(n log n) per group. Select only when you need true distributional statistics.
Honest Limitations: Where Falls Short
No tool is perfect for every use case. Here's where might be a better choice, and the real limitations of our browser-based architecture.
Browser-Based Processing
Performance depends on your device's RAM and CPU. Modern laptops (2022+) handle 10M+ rows easily, but older devices may struggle with very large files.
No Offline Mode (Initial Load)
Requires internet connection to load the tool initially. Processing happens offline in your browser after loading.
Browser Tab Memory Limits
Most browsers limit individual tabs to 2-4GB RAM. This is the practical ceiling for file size.
Questions about limitations? Check our FAQ section below or contact us via the feedback button.
Frequently Asked Questions
How accurate are the 34-second and 26-second benchmarks for 10M rows?
Why is advanced mode slower than simple mode?
Why does performance improve with larger files (1K rows is slower than 10M)?
How does this compare to Excel Pivot Tables at scale?
What happens with high-cardinality GROUP BY columns?
Does the HAVING filter affect processing speed?
What's the maximum file size SplitForge can handle?
What browser gives the best performance?
Does SplitForge transmit any file data for analytics?
Benchmarks last updated: February 2026 · Conducted by the SplitForge engineering team · Re-tested quarterly and after major algorithm changes · Aggregation Engine v1.0 · SVBP-2026 Protocol (SplitForge Verified Benchmark Protocol: synthetic dataset, 10-run median, cold-start excluded, hardware-documented) · Privacy verified: no outbound file payload requests (network inspector confirmed)