SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 19 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Paradigm Shift
Achieves 'zero forgetting' in continual learning by stacking frozen domain-specific MoE-LoRA adapters with a meta-router.
Apr 2
New Capability
First humanoid robot system to achieve consecutive ping-pong strikes using only onboard egocentric vision and whole-body coordination.
Apr 2
Breaks Assumption
Reveals a 'Reasoning Shift' where increased context length silently causes models to skip self-verification and shorten their reasoning traces by up to 50%.
Apr 2
Efficiency Breakthrough
Introduces S0 tuning for hybrid RNN-attention models, outperforming LoRA by 10.8% with zero inference overhead.
Apr 2
Efficiency Breakthrough
Reduces the compute cost of LLM test-time scaling by up to 67% using conformal prediction to calibrate reasoning paths.
Apr 2
Paradigm Shift
Replaces standard relative Softmax attention with 'Multiscreening' to allow absolute query-key relevance, yielding 3.2x faster inference at 100K context.
Apr 2
Scaling Insight
Simple Self-Distillation (SSD) improves LLM code generation (e.g., Qwen3-30B) by 13% Pass@1 without any external verifiers or teacher models.
Apr 2
Breaks Assumption
Provides causal evidence that reasoning models often decide on an action (like a tool call) before they even start generating their 'Chain-of-Thought'.
Apr 2
Efficiency Breakthrough
Combines the YOCO architecture with recursive computation to scale representational depth without inflating the KV cache.
Apr 2
Efficiency Breakthrough
Solves the long-standing trade-off in low-rank matrix recovery by achieving both optimal sample complexity and fast convergence.
Apr 2
Breaks Assumption
Provides a theoretical explanation for why Transformers often fail compared to linear models in financial time series forecasting.
Apr 2
Efficiency Breakthrough
Enables Gaussian Processes to scale on modern parallel hardware by removing the need for Cholesky decompositions.
Apr 2
New Capability
Introduces 'deconfounding scores' to enable reliable causal effect estimation even when treatment and control groups have very little overlap.
Apr 2
Open Release
Delivers a state-of-the-art universal phone recognition model across 100+ languages with full open-source release.
Apr 2
Cosmic Scale
Researchers have designed a new internet protocol specifically for a 10-node colony network spanning Earth, the Moon, and Mars.
Apr 1
Practical Magic
Everyday 5G cell towers can be repurposed as a massive radar system capable of tracking drones hidden in urban noise.
Apr 1
Nature Is Weird
AI voice assistants can be tricked into 'hearing' voices and events that never actually happened with near-perfect accuracy.
Apr 1
Practical Magic
Future wireless signals could be boosted by walls that physically shift and morph their shape to bounce waves toward your phone.
Apr 1
Paradigm Challenge
Researchers have mapped out all 19.3 million chords the human hand can play on a piano to reveal why some sound 'clear' and others 'muddy.'
Apr 1
New Capability
Interfaces LLMs with Wikidata-scale graphs for multi-hop reasoning without any retraining of the model or the query executor.
Apr 1
Open Release
A unified, open-source framework that converts complex post-training quantization workflows into a single-line, hardware-aware pipeline.
Apr 1
Efficiency Breakthrough
Decouples data mixture ratio selection from continual pre-training by optimizing distribution vectors post-hoc with 15-35x lower compute cost.
Apr 1
New Capability
Achieves an 80x improvement in stable generation length for occupancy world models, enabling 4km+ autonomous driving simulations from a single frame.
Apr 1
Paradigm Shift
Replaces the heuristic constant momentum (0.9) with a parameter-free, physics-inspired schedule that speeds up convergence by nearly 2x.
Apr 1
New Capability
Leverages model reprogramming as an 'active signal amplifier' to proactively audit privacy leakage in LLMs and Diffusion models.
Apr 1
Efficiency Breakthrough
Combines differentiable optimization with exact ILP solvers to achieve a 10x performance gain in solving NP-hard combinatorial scheduling problems.
Apr 1
Paradigm Shift
Proposes a mathematical framework where 'spectral gaps' in parameter updates control phase transitions like grokking and loss plateaus.
Apr 1
Breaks Assumption
Large-scale experiments reveal that self-organizing LLM agents spontaneously outperform manually designed hierarchical structures by 14%.
Apr 1
Efficiency Breakthrough
A fabricated 16nm SoC that performs real-time 3D occupancy mapping under 6 mW, reducing query energy by over 80%.
Apr 1
Paradigm Shift
Proposes a neuroscience-grounded memory architecture that makes interactions cheaper and more accurate with experience, rather than relying on expanding context windows.
Apr 1
Breaks Assumption
Reveals that parallel translated data is surprisingly unnecessary for creating aligned multilingual representations in LLMs.
Apr 1
Breaks Assumption
Discovers that pretraining Implicit Neural Representations (INRs) on structured $1/f^\alpha$ noise performs as well as data-driven initialization.
Apr 1
Paradigm Shift
Introduces DASES, a framework that replaces passive validation with active 'falsification' to ensure scientific models learn actual mechanisms rather than just winning benchmarks.
Apr 1
Efficiency Breakthrough
Generates complete, simulatable analog circuits in milliseconds, outperforming search-based methods by over 600x.
Apr 1
Breaks Assumption
Demonstrates that integer multiplication is not a long-range dependency problem, and that current architectures like Transformers and Mamba are fundamentally using the wrong 'computational spacetime.'
Apr 1
Efficiency Breakthrough
Introduces PolarQuant, a quantization method that uses Hadamard rotation to make LLM weights near-lossless at 5-bit without calibration data.
Apr 1
Breaks Assumption
Demonstrates that the 'modality gap' in CLIP-style models is a feature that can be exploited to increase robustness without retraining.
Apr 1
New Capability
Achieves a +48pp accuracy gain in agents using a non-parametric online learning framework that reuses procedural plans without updating model weights.
Apr 1
Efficiency Breakthrough
Scales curvature-aware bilevel optimization to BERT-sized models using KFAC, significantly outperforming standard gradient unrolling.
Apr 1
Paradigm Shift
Switches the training objective from hard Next-Token Prediction to predicting 'concepts' (sets of semantically related tokens).
Apr 1
Breaks Assumption
Challenges the assumption that architecture and loss are the primary levers for neural simulators by proving the 'carried state' design is the dominant bottleneck.
Apr 1
Paradigm Shift
Proves that LLM agent capability (pass@1) and reliability (consistency) diverge systematically, with frontier models often having the highest 'meltdown' rates.
Apr 1
New Capability
Introduces a way for diffusion models to generate a single, sharp 'mental average' of a concept rather than blurry pixel-wise averages.
Apr 1
Open Release
A massive multimodal release for 10 low-resource African languages, reducing SOTA Word Error Rates (WER) by up to 61% relative.
Apr 1
Efficiency Breakthrough
Enables infinite-length video understanding on a single consumer GPU (RTX 3090) through a training-free visual memory mechanism.
Apr 1
Paradigm Shift
Learns stable, interpretable Koopman generators for nonlinear PDEs from trajectory data alone without any physics supervision.
Apr 1
Open Release
A massive 270K-sample multi-view video corpus specifically for embodied AI agents in complex retail environments.
Apr 1
New Capability
Introduces a scalable reinforcement learning framework that enables high-fidelity control of a whole-body human musculoskeletal system with over 700 muscles.
Apr 1
New Capability
Proposes 'Nomad', an exploration-first agent architecture that autonomously discovers insights in data without being limited by human prompts or questions.
Apr 1
Breaks Assumption
Reveals that many massive LLM benchmarks provide highly redundant information, with major leaderboards often containing only ~2 independent axes of measurement.
Apr 1