Neural Networks

ARC-AGI Without Pretraining

A novel approach demonstrates that lossless information compression during inference time can produce intelligent behavior, achieving 34.75% accuracy on ARC-AGI training set without pretraining or extensive datasets. The method, CompressARC, processes each puzzle in 20 minutes using only compression objectives and efficient inference-time computation, challenging conventional reliance on extensive pretraining and data.

Crossing the uncanny valley of conversational voice

Sesame introduces Conversational Speech Model (CSM), advancing voice AI beyond traditional text-to-speech limitations by incorporating contextual awareness and emotional intelligence. The model operates as a single-stage system using transformers to produce more natural and coherent speech, achieving near-human performance in audio quality while still working to improve conversational dynamics.

The FFT Strikes Back: An Efficient Alternative to Self-Attention

FFTNet introduces a novel approach to sequence processing using Fast Fourier Transform, achieving O(n log n) complexity compared to traditional self-attention's quadratic complexity. The framework employs spectral filtering and modReLU activation to efficiently capture long-range dependencies, demonstrating superior performance on Long Range Arena and ImageNet benchmarks.

Helix: A Vision-Language-Action Model for Generalist Humanoid Control

Figure introduces Helix, a groundbreaking Vision-Language-Action model capable of controlling humanoid robot upper bodies through natural language commands. The system uniquely combines high-speed continuous control with multi-robot collaboration capabilities, operating entirely on embedded GPUs. Helix demonstrates remarkable ability to manipulate thousands of novel objects without prior training, marking a significant advancement in scalable robotics.

Ask HN: Is anybody building an alternative transformer?

Various alternative architectures to Transformers are being explored, with MAMBA showing promise through faster inference and lower compute costs, performing on par with transformers up to 7B parameters. Researchers are investigating recurrent architectures, state-space models, and efficient attention mechanisms, while debating the future direction of foundation models.

Automating GPU Kernel Generation with DeepSeek-R1 and Inference Time Scaling | NVIDIA Technical Blog

NVIDIA engineers utilized the DeepSeek-R1 model with inference-time scaling to automatically generate optimized GPU attention kernels, achieving results that sometimes surpassed human-engineered solutions. The experiment demonstrates how AI models can leverage additional computational resources during inference to evaluate multiple outcomes and select optimal solutions for complex programming tasks.

Les enjeux de l’IA : mon interview sur France 2 et Firstpost.

Transformers' extraordinary learning capabilities allow them to master skills through simple observation of related tasks, showcasing the potential of emergent behavior in AI. Recent studies demonstrate that transformer models can learn complex skills without explicit training, revealing profound implications for future AI development and understanding.

Deep dive into LLMs like ChatGPT by Andrej Karpathy (TL;DR)

Andrej Karpathy's deep dive into LLMs covers the complete lifecycle from pretraining to post-training, explaining tokenization, neural network architectures, and fine-tuning processes. The comprehensive guide explores how LLMs process information, handle hallucinations, and utilize reinforcement learning to improve performance and reasoning capabilities.

GitHub - agentsea/r1-computer-use: Applying the ideas of Deepseek R1 to computer use

An experimental project applying large-scale Reinforcement Learning techniques to computer usage scenarios, utilizing neural reward models to validate agent actions. The system implements a three-step cycle extending ReACT into reinforcement learning, with multiple training stages focused on developing reasoning skills for computer interaction.