AI pulse last 7 days
Daily AI pulse from YouTube, blogs, Reddit, HN. Ruthlessly filtered.
Sources (41)▶
- criticalAndrej Karpathy
Były dyrektor AI w Tesli, OpenAI cofounder. Każde video to gold.
- criticalAnthropic
Oficjalny kanał Anthropic. Każdy release Claude'a.
- criticalComfyUI Blog
Release log dla integracji ComfyUI — Luma Uni-1, GPT Image 2, ACE-Step music gen, Seedance. Pokrywa video+image+music+workflow.
- criticalOpenAI Blog
Oficjalny blog OpenAI. Wszystkie release.
- criticalSimon Willison's Weblog
Najlepszy 'thinker' AI. Codzienne posty, deep insights, niska hype rate.
- highAI Explained
Głęboka analiza papers i benchmarków, niska hype rate.
- highAI Jason
Praktyczne tutoriale Claude Code, MCP, workflow vibe codingu.
- highBen's Bites
Daily AI digest, creator-friendly tone. Codex, model releases, agentic AI.
- highCole Medin
Vibe coding + agentic workflows + Claude Code MCP integrations.
- highFal AI Blog
Fal hostuje większość nowych AI image/video modeli — ich blog to wczesne sygnały premier.
- highHN: 3D & Gaussian Splatting
HN signal dla 3D generative — Gaussian Splatting, NeRF, image-to-3D. Próg 20 bo niszowa kategoria (top historic 182pts).
- highHN: AI agents / MCP
HN posty o agentach, MCP, vibe codingu z min 100 pkt.
- highHN: Claude / Anthropic
HN posty z 'Claude' lub 'Anthropic' z min 100 pkt.
- highHugging Face Blog
Releases dla image, video, audio, 3D modeli. Część tech-heavy — Gemini relevance odfiltruje noise. Downgraded z critical: za duży volume na 'must-read' status.
- highIndyDevDan
Claude Code power user, prompty, hooki.
- highInterconnects (Nathan Lambert)
AI policy + research analysis. Niska hype rate, opinionated.
- highLatent Space
Podcast + blog Swyx — wywiady z founderami i deep dives engineeringowe.
- highMatt Wolfe
Comprehensive AI tools weekly digest. ~700K subs.
- highMatthew Berman
AI news, model release reviews, agent demos. Wysoki output.
- highr/aivideo
Community AI video — Sora, Veo, Runway, Kling, LTX. Co naprawdę zaskakuje twórców.
- highr/ClaudeAI
Społeczność Claude'a — power users, tipy, problemy.
- highr/LocalLLaMA
Open-source LLMs, lokalne uruchamianie, benchmarks bez hype.
- highr/StableDiffusion
Największa community open-source image gen (700k+ users). Premiery modeli, LoRA, ComfyUI workflows.
- highRiley Brown
Vibe coding, AI builder workflows, Cursor + Claude tutorials.
- highThe Decoder
Niemiecki AI news outlet po angielsku, dobre breaking news.
- highTheo - t3.gg
TypeScript + AI dev workflows. Hot takes, narrative-driven.
- highYannic Kilcher
Paper reviews i deep dives w research AI.
- lowAI Weirdness
Janelle Shane — playful AI experiments, image gen quirks. Niski volume, unikalna perspektywa.
- mediumbycloud
AI papers digestible — między 2MP a Yannic Kilcher.
- mediumCreative Bloq
Design industry — gdzie AI ingeruje w klasyczne dyscypliny graficzne.
- mediumFireship
100-sec format, often AI/LLM + tech news.
- mediumfxguide
VFX i film industry — coraz więcej AI w pipeline. Profesjonalna perspektywa.
- mediumGreg Isenberg
Solo founder vibe — buduje produkty z AI, podcasty z indie hackers.
- mediumr/ChatGPTCoding
Vibe coding tipy, IDE setupy, prompty. Mix wszystkich modeli.
- mediumr/comfyui
ComfyUI workflows — custom nodes, JSON workflows, optymalizacje.
- mediumr/midjourney
Midjourney community — premiery v7+, style references, prompt patterns.
- mediumr/runwayml
Runway-specific community — premiery features, prompt patterns, comparisons z konkurencją.
- mediumr/SunoAI
Suno music gen community — nowe wersje modelu, lyric prompting techniques. Audio AI ma slaby RSS ecosystem.
- mediumTina Huang
AI workflows for data science, practical applications.
- mediumTwo Minute Papers
Krótkie streszczenia papers AI, świetne dla szybkiego scan'a.
- mediumWes Roth
AI news z bardziej clickbaitowym tonem — filtr Gemini odsiewa hype.

Google speeds up Gemma 4 threefold with multi-token prediction
You can now generate text with Google's Gemma 4 models up to three times faster thanks to a new multi-token prediction technique.
Google has introduced multi-token prediction drafters for its Gemma 4 open model family, significantly accelerating text generation. This new feature allows Gemma 4 models to generate text up to three times faster than before. The technique involves a smaller auxiliary model that proposes several tokens simultaneously, which the main Gemma model then validates in a single pass. This enhancement provides a substantial performance boost for users working with Gemma 4, making it more efficient for various creative and development tasks.
The Decoder·model_release·05/06/2026, 04:05 PM·Matthias Bastian
Google and Meta race to build personal AI agents as Anthropic and OpenAI pull further ahead
Google and Meta are pivoting from browser-based automation to deeply integrated personal agents (Remy and Hatch) to compete with OpenAI and Anthropic.
Google and Meta are intensifying their efforts to develop autonomous personal AI agents, codenamed "Remy" and "Hatch" respectively. This move is a strategic pivot to counter the early lead established by OpenAI and Anthropic in the agentic space. Notably, Google has reportedly halted its "Mariner" browser agent project to consolidate resources into these more integrated solutions. The industry trend is moving away from agents that simply control a web browser toward assistants embedded directly into core services like email, calendars, and e-commerce. These new agents aim to handle complex, multi-step everyday tasks autonomously within the platforms users already inhabit.
The Decoder·news·05/06/2026, 12:53 PM·Maximilian Schreiner
US and tech firms strike deal to review AI models for national security before public release | Technology
The US government and major AI labs have agreed on a pre-release review process for new models to assess national security risks before they go public.
The US Department of Commerce has reached a landmark agreement with leading AI companies, including Google, Microsoft, and xAI, to implement a pre-release review system for advanced AI models. This voluntary framework allows government agencies to evaluate potential national security risks, such as cyber-warfare capabilities or biological threat assistance, before public deployment. The move signals a shift towards tighter oversight of frontier LLM development as concerns over existential risks and global competition intensify. While the agreement is currently voluntary, it sets a precedent for how future high-capability models might be regulated. For the open-source community, this raises questions about whether similar scrutiny will eventually extend to weights shared on platforms like…
r/LocalLLaMA·news·05/05/2026, 09:46 PM·/u/Merchant_LawrenceGemma 4 MTP released
Google released MTP draft models for Gemma 4, enabling up to 2x faster generation through speculative decoding without sacrificing output quality.
Google has officially released Multi-Token Prediction (MTP) draft models for the Gemma 4 family, including the 31B and various MoE variants. MTP works by pairing the base model with a smaller, faster draft model that predicts multiple tokens ahead. These predictions are then verified in parallel by the main model using a Speculative Decoding pipeline. This approach achieves up to a 2x speedup in inference speed, which is critical for local and on-device deployments. Crucially, the final output remains identical to standard generation, offering a significant performance boost for supported hardware and software stacks without sacrificing quality.
r/LocalLLaMA·model_release·05/05/2026, 04:01 PM·/u/rerri
Gemma 4 MTP released
Get up to 2x faster inference on Gemma 4 models using the newly released Multi-Token Prediction draft checkpoints for speculative decoding.
Google has officially released Multi-Token Prediction (MTP) draft models for the Gemma 4 family, including variants for the 31B and smaller models. These draft models are designed for Speculative Decoding, where a smaller model predicts multiple future tokens that the main model then validates in parallel. This technique can achieve up to 2x speedups in generation latency while maintaining identical output quality compared to standard autoregressive generation. The release includes specialized checkpoints on Hugging Face tuned as assistants for the main Gemma 4 weights. This is a significant update for local LLM users and on-device applications where inference speed is often the primary bottleneck.
r/LocalLLaMA·model_release·05/05/2026, 04:01 PM·rerri
Supercharging LLM inference on Google TPUs: Achieving 3X speedups with diffusion-style speculative decoding- Google Developers Blog
Google achieved a 3X speedup in LLM inference on TPUs by using a new 'diffusion-style' parallel token drafting technique.
Google researchers have introduced a novel approach to speculative decoding inspired by diffusion models, specifically optimized for TPU architectures. Traditional speculative decoding relies on a smaller draft model to predict tokens sequentially, but this new method generates multiple draft tokens in parallel, similar to how diffusion models refine images. This shift addresses the memory bandwidth bottlenecks common in LLM inference, resulting in up to 3X faster generation speeds. While the benchmarks focus on Google's proprietary hardware, the move toward non-autoregressive drafting represents a significant evolution in inference strategy. This technique could eventually influence local model optimization if adapted for consumer GPUs.
r/LocalLLaMA·news·05/05/2026, 03:50 PM·/u/eternviking
Gemini 3.5 Flash In Arena! POWERFUL, Cheap, & Fast NEW AI Model! (Fully Tested)
A new, significantly more powerful Gemini Flash variant is being tested in the Arena, offering Pro-level coding and reasoning performance at a 'Flash' price point.
Google is quietly testing an upgraded Gemini Flash model in the LMSYS Chatbot Arena, likely as a precursor to a major Google I/O announcement. Early testers report a massive leap in reasoning and output quality, with performance reportedly rivaling the more expensive Gemini 3.1 Pro. The video showcases the model's proficiency in complex front-end coding tasks using React and GSAP, as well as 3D scene generation in Three.js, such as a functional 1970s TV simulator. While the exact version number (3.1 vs 3.5) remains unconfirmed, the update signals a significant shift in the price-to-performance ratio for Google's lightweight model tier.
AI Jason·news·05/03/2026, 06:44 AM·WorldofAI▶Watch here
Relevance auto-scored by LLM (0–10). List shows top 30 from the last 7 days.