About:

Grigory Sapunov is a Co-Founder and CTO at Intento, with a focus on ML and AI insights.

Website:

Specializations:

Interests:

Machine Learning AI-generated insights Expert curation

Outgoing Links:

Grigory Sapunov
Subscribe to RSS:
The authors present a novel pipeline that generates unconventional research ideas by decomposing machine learning papers into 'idea atoms' and training models to explore non-obvious directions.
Solaris is a groundbreaking multi-agent video world model for Minecraft that enables consistent multi-view observations and addresses the limitations of single-agent architectures.
An innovative algorithmic co-design for attention computation on NVIDIA's Blackwell architecture enhances performance by addressing hardware scaling challenges, achieving up to 1613 TFLOPs/s.
The authors propose Superhuman Adaptable Intelligence (SAI) as a more effective framework than AGI, emphasizing specialization and adaptability in AI systems.
The study reveals that repository-level context files for coding agents can hinder performance and increase costs, challenging their recommended use in software engineering.
Semantic Tube Prediction enhances LLM training efficiency by constraining hidden states to smooth trajectories, achieving high accuracy with significantly less data.
Vox Deorum presents a hybrid AI architecture that enhances gameplay in Civilization V by decoupling high-level strategy from tactical execution, achieving competitive performance metrics.
Causal-JEPA enhances object-centric world models by using innovative masking techniques to improve interaction reasoning and computational efficiency in model predictive control tasks.
This study exposes significant vulnerabilities in autonomous language-model agents, highlighting the urgent need for improved safety and governance in AI deployments.
The Deep-Thinking Ratio (DTR) quantifies reasoning effort in language models, offering a more efficient alternative to traditional token count metrics for improving inference accuracy.
A new framework models Long Chain-of-Thought reasoning as a molecular structure, emphasizing topological distributions and introducing MOLE-SYN to enhance weaker instruction models.
A new reinforcement learning framework allows a Large Language Model to autonomously optimize CUDA kernels, surpassing traditional compiler methods and enhancing deep learning performance.
The study reveals that unified multimodal pretraining can enhance AI capabilities by integrating language and vision without relying on text-heavy models.
The study reveals that dense MLP layers in Large Language Models inherently perform sparse computations similar to Mixture of Experts layers, bridging theory and empirical design.
Speculative Speculative Decoding (SSD) optimizes language model decoding by enabling parallel processing of draft predictions and verification, resulting in substantial speed improvements.
Memory Caching enhances RNNs by allowing them to cache memory states, improving recall performance while maintaining computational efficiency compared to Transformers.
A unified theory reveals that geometric representations in language models emerge from translation symmetry in co-occurrence statistics, challenging assumptions about complex learning dynamics.
The study reveals that Transformers are significantly less data-efficient than RNNs for state-tracking tasks, requiring exponentially more data for convergence.
AlphaEvolve leverages LLMs to automatically create novel algorithms for Multi-Agent Reinforcement Learning, surpassing human-designed methods in effectiveness.
Unified Latents (UL) optimizes generative modeling by linking noise in latent space to diffusion model precision, enhancing efficiency and performance on key datasets.
CogRouter introduces a framework for dynamically modulating cognitive depth in large language models, improving efficiency and performance in long-horizon tasks.
The 'Theory of Space' framework evaluates MLLMs on their ability to actively explore environments and construct spatial beliefs, revealing critical gaps in current model capabilities.
A new economic framework reveals that the transition to AGI poses systemic risks due to the imbalance between automation costs and human verification capacity.
An analysis of AI agents on Moltbook shows rapid hierarchical organization and attention inequality, raising concerns about systemic risks in multi-agent ecosystems.