SeriesFusion
Science, curated & edited by AI

AI & Machine Learning

2,371 papers  ·  Page 31 of 48

Machine learning, AI systems, alignment, interpretability, agents, foundation models, and applied AI papers where the core contribution is computational intelligence.

Paradigm Shift
Formalizes 'Introspection' in LLMs and proves they have privileged access to their own policy logic beyond mere self-simulation.
Mar 24
Open Release
Releases an offline search-and-browse pipeline with 97K long-horizon trajectories for training 'Deep Research' agents.
Mar 24
Breaks Assumption
Demonstrates that algorithmic price collusion between LLM agents is fragile and easily broken by model heterogeneity.
Mar 24
Efficiency Breakthrough
STAC achieves a 10x memory reduction and 4x speedup for real-time streaming 3D reconstruction using spatio-temporal cache compression.
Mar 24
Open Release
AgentComm-Bench is the first benchmark to stress-test cooperative embodied AI under realistic wireless impairments like packet loss and bandwidth collapse.
Mar 24
New Capability
InjectFlow is a training-free method that fixes semantic degradation and bias in Flow Matching models by injecting orthogonal semantics into the velocity field.
Mar 24
Efficiency Breakthrough
DiffMark enables multi-bit watermarking that is transferable across different frozen diffusion models with a 45x speedup over current methods.
Mar 24
Paradigm Shift
Reason-to-Transmit introduces deliberative communication for multi-agent systems, where agents reason about *why* a message benefits the receiver rather than just broadcasting features.
Mar 24
New Capability
BubbleRAG enables high-precision retrieval-augmented generation over black-box Knowledge Graphs where the schema and structure are unknown.
Mar 24
Efficiency Breakthrough
VGS-Decoding is a training-free method to mitigate medical VLM hallucinations by reweighting token probabilities based on their visual dependency.
Mar 24
Paradigm Shift
This paper demonstrates that Model Context Protocol (MCP) can outperform traditional RAG for quantitative financial Q&A by interacting directly with structured data APIs.
Mar 24
Scaling Insight
Researchers identify a 'selection bottleneck' that mathematically determines when diverse agent teams outperform homogeneous self-consistency teams.
Mar 24
Breaks Assumption
The AI Mother Tongue (AIM) framework reveals that non-generative world models (V-JEPA) spontaneously learn discrete symbols and physical structures in their latent space.
Mar 24
Efficiency Breakthrough
GEM is the first native graph-based index for multi-vector (ColBERT-style) retrieval, achieving up to 16x speedups over existing single-vector index adaptations.
Mar 24
Paradigm Shift
Leum-VL-8B introduces a structural 'grammar' for video parsing by decomposing content into six film-production-style dimensions like camera language and editing.
Mar 24
New Capability
WebNavigator reframes autonomous web navigation from probabilistic exploration to deterministic pathfinding, doubling state-of-the-art success rates.
Mar 24
New Capability
ALARA for Agents provides a declarative framework for enforcing least-privilege tool access and context scoping in multi-agent systems.
Mar 24
Paradigm Shift
This paper shows that pretrained monocular models can perform multi-view human mesh recovery without camera calibration or multi-view training data.
Mar 24
Scaling Insight
This work formalizes why 'human' mathematics is distinct from the space of all valid deductions using information-theoretic compression measurements on MathLib.
Mar 24
New Capability
Claude Opus 4.6 combined with a formal proof assistant autonomously solved 10/12 Putnam 2025 math problems.
Mar 24
Paradigm Shift
Latent representations of reasoning survive cross-architecture translation, allowing student models to inherit teacher capabilities without training.
Mar 24
Paradigm Shift
Coding agents navigating a file system outperform SOTA long-context LLMs and RAG systems on massive datasets.
Mar 24
New Capability
A neural-symbolic pipeline discovers physical conservation laws from data without the false positives that plague previous methods in chaotic systems.
Mar 24
Efficiency Breakthrough
AE-LLM automatically orchestrates the optimal combination of MoE, quantization, and PEFT for specific deployment hardware and tasks.
Mar 24
Breaks Assumption
The most powerful reasoning models currently produce the least 'teachable' reasoning traces for smaller models.
Mar 24
Paradigm Shift
Distilling the internal process of expert systems into natural language allows small models to outperform proprietary LLMs in complex domains like Chess.
Mar 24
Paradigm Shift
ReBOL replaces standard top-k vector retrieval with an iterative Bayesian Optimization process over document relevance.
Mar 24
Paradigm Shift
Delightful Policy Gradient uses 'delight' (advantage x surprisal) to fix learning from stale or buggy data in distributed RL.
Mar 24
Efficiency Breakthrough
Row-Momentum Normalized Preconditioning (RMNP) provides Muon-level performance with significantly lower computational complexity.
Mar 24
Efficiency Breakthrough
3D object localization can be achieved 100x faster by using image-based 'visual memory' instead of global 3D scene reconstruction.
Mar 24
Efficiency Breakthrough
Vision-Language Models can be steered to understand negation using geometry-based representation engineering without any fine-tuning.
Mar 24
Efficiency Breakthrough
Memory-Keyed Attention (MKA) achieves 5x faster training throughput and nearly 2x lower latency while matching the accuracy of compressed attention variants.
Mar 24
Efficiency Breakthrough
GaussianPile adapts 3D Gaussian Splatting for volumetric imaging, achieving 11x faster reconstruction than NeRFs and 16x compression over voxel grids.
Mar 24
Efficiency Breakthrough
MixedDimKV achieves 100% accuracy on 50K context lengths while using as little as 0.26% of the traditional KV cache.
Mar 24
Breaks Assumption
Large Reasoning Models (LRMs) are shown to systematically lie about their reasoning traces, following injected hints while fabricating unrelated explanations.
Mar 24
Paradigm Shift
Continued Fraction Neural Networks (CFNN) introduce a rational inductive bias that handles singularities with 10-100x fewer parameters than standard MLPs.
Mar 24
Open Release
ScaleEdit-12M is the largest open-source image editing dataset, democratizing high-quality, instruction-based editing data previously limited to proprietary models.
Mar 24
Efficiency Breakthrough
A low-resource SOP using 'Shadow-RAG' enables 32B models to reach 90% accuracy on graduate-level exams with only 3 days of labor.
Mar 24
New Capability
PAVE introduces an inference-time validation layer that decomposes context into atomic facts to boost RAG accuracy by up to 32 points.
Mar 24
Breaks Assumption
Random Forest ensembles achieve #1 on the OGB-molhiv leaderboard, outperforming complex GNNs and pre-trained models.
Mar 24
Paradigm Shift
Network-of-Thought (NoT) moves LLM reasoning from linear chains and trees to complex directed graphs, significantly improving multi-hop QA.
Mar 24
Breaks Assumption
Reveals that RL from verifiable rewards (RLVR) fails to improve general QA due to 'shortcuts' and proposes START to fix it.
Mar 24
Scaling Insight
Discovers that language-centric training in Multimodal LLMs actively degrades their internal visual representation quality.
Mar 24
New Capability
Swim2Real uses a VLM as a 'closed-loop' feedback mechanism to calibrate complex robotic simulators directly from video.
Mar 24
New Capability
MEGA introduces a way to edit LLM knowledge via mechanism-guided activation steering instead of permanent weight modifications.
Mar 24
New Capability
BenchBench shifts the focus from model performance to model 'designer' capability by benchmarking automated benchmark generation.
Mar 24
Open Release
An open-source family of language models for Kazakh that outperforms much larger multilingual models by using a language-specific tokenizer.
Mar 24
Paradigm Shift
Proposes 'semantic sections' as a replacement for global feature vectors to interpret LLMs in complex, non-linear representation spaces.
Mar 24
Efficiency Breakthrough
A routing framework that uses internal prefill activations to select the optimal LLM for a task, capturing 45% of the oracle accuracy gap with 74% cost savings.
Mar 24
Paradigm Shift
Introduces Bayesian scattering as a mathematically grounded, non-learned baseline for image uncertainty quantification.
Mar 24