Skip to content

zetabyte

Flow State to Free Fall: An AI Coding Cautionary Tale Sreeram Venkatasubramanian AI & ML – Radar

​[[{“value”:” When I was eight years old, I watched a mountaineering documentary while waiting for the cricket match to start. I remember being incredibly frustrated watching these climbers inch their way up a massive rock face, stopping every few feet to hammer what looked like… Read More »Flow State to Free Fall: An AI Coding Cautionary Tale Sreeram Venkatasubramanian AI & ML – Radar

StreamBridge: Turning Your Offline Video Large Language Model into a Proactive Streaming Assistant Apple Machine Learning Research

​We present StreamBridge, a simple yet effective framework that seamlessly transforms offline Video-LLMs into streaming-capable models. It addresses two fundamental challenges in adapting existing models into online scenarios: (1) limited capability for multi-turn real-time understanding, and (2) lack of proactive response mechanisms. Specifically, StreamBridge incorporates… Read More »StreamBridge: Turning Your Offline Video Large Language Model into a Proactive Streaming Assistant Apple Machine Learning Research

Checklists Are Better Than Reward Models For Aligning Language Models Apple Machine Learning Research

​Language models must be adapted to understand and follow user instructions. Reinforcement learning is widely used to facilitate this — typically using fixed criteria such as “helpfulness” and “harmfulness”. In our work, we instead propose using flexible, instruction-specific criteria as a means of broadening the… Read More »Checklists Are Better Than Reward Models For Aligning Language Models Apple Machine Learning Research

The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity Apple Machine Learning Research

​Recent generations of frontier language models have introduced Large Reasoning Models (LRMs) that generate detailed thinking processes before providing answers. While these models demonstrate improved performance on reasoning benchmarks, their fundamental capabilities, scaling properties, and limitations remain insufficiently understood. Current evaluations primarily focus on established… Read More »The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity Apple Machine Learning Research

Enabling Differentially Private Federated Learning for Speech Recognition: Benchmarks, Adaptive Optimizers, and Gradient Clipping Apple Machine Learning Research

​While federated learning (FL) and differential privacy (DP) have been extensively studied, their application to automatic speech recognition (ASR) remains largely unexplored due to the challenges in training large transformer models. Specifically, large models further exacerbate issues in FL as they are particularly susceptible to… Read More »Enabling Differentially Private Federated Learning for Speech Recognition: Benchmarks, Adaptive Optimizers, and Gradient Clipping Apple Machine Learning Research

PREAMBLE: Private and Efficient Aggregation via Block Sparse Vectors Apple Machine Learning Research

​[[{“value”:”We revisit the problem of secure aggregation of high-dimensional vectors in a two-server system such as Prio. These systems are typically used to aggregate vectors such as gradients in private federated learning, where the aggregate itself is protected via noise addition to ensure differential privacy.… Read More »PREAMBLE: Private and Efficient Aggregation via Block Sparse Vectors Apple Machine Learning Research

Generative and Contrastive Graph Representation Learning Apple Machine Learning Research

​Self-supervised learning (SSL) on graphs generates node and graph representations (i.e., embeddings) that can be used for downstream tasks such as node classification, node clustering, and link prediction. Graph SSL is particularly useful in scenarios with limited or no labeled data. Existing SSL methods predominantly… Read More »Generative and Contrastive Graph Representation Learning Apple Machine Learning Research

Gemini Robotics 1.5: DeepMind’s ER↔VLA Stack Brings Agentic Robots to the Real World Asif Razzaq Artificial Intelligence Category – MarkTechPost

​[[{“value”:” Can a single AI stack plan like a researcher, reason over scenes, and transfer motions across different robots—without retraining from scratch? Google DeepMind’s Gemini Robotics 1.5 says yes, by splitting embodied intelligence into two models: Gemini Robotics-ER 1.5 for high-level embodied reasoning (spatial understanding,… Read More »Gemini Robotics 1.5: DeepMind’s ER↔VLA Stack Brings Agentic Robots to the Real World Asif Razzaq Artificial Intelligence Category – MarkTechPost

Top 10 Local LLMs (2025): Context Windows, VRAM Targets, and Licenses Compared Michal Sutter Artificial Intelligence Category – MarkTechPost

​[[{“value”:” Local LLMs matured fast in 2025: open-weight families like Llama 3.1 (128K context length (ctx)), Qwen3 (Apache-2.0, dense + MoE), Gemma 2 (9B/27B, 8K ctx), Mixtral 8×7B (Apache-2.0 SMoE), and Phi-4-mini (3.8B, 128K ctx) now ship reliable specs and first-class local runners (GGUF/llama.cpp, LM… Read More »Top 10 Local LLMs (2025): Context Windows, VRAM Targets, and Licenses Compared Michal Sutter Artificial Intelligence Category – MarkTechPost