AI pulse last 7 days
Daily AI pulse from YouTube, blogs, Reddit, HN. Ruthlessly filtered.
Sources (41)▶
- criticalAndrej Karpathy
Były dyrektor AI w Tesli, OpenAI cofounder. Każde video to gold.
- criticalAnthropic
Oficjalny kanał Anthropic. Każdy release Claude'a.
- criticalComfyUI Blog
Release log dla integracji ComfyUI — Luma Uni-1, GPT Image 2, ACE-Step music gen, Seedance. Pokrywa video+image+music+workflow.
- criticalOpenAI Blog
Oficjalny blog OpenAI. Wszystkie release.
- criticalSimon Willison's Weblog
Najlepszy 'thinker' AI. Codzienne posty, deep insights, niska hype rate.
- highAI Explained
Głęboka analiza papers i benchmarków, niska hype rate.
- highAI Jason
Praktyczne tutoriale Claude Code, MCP, workflow vibe codingu.
- highBen's Bites
Daily AI digest, creator-friendly tone. Codex, model releases, agentic AI.
- highCole Medin
Vibe coding + agentic workflows + Claude Code MCP integrations.
- highFal AI Blog
Fal hostuje większość nowych AI image/video modeli — ich blog to wczesne sygnały premier.
- highHN: 3D & Gaussian Splatting
HN signal dla 3D generative — Gaussian Splatting, NeRF, image-to-3D. Próg 20 bo niszowa kategoria (top historic 182pts).
- highHN: AI agents / MCP
HN posty o agentach, MCP, vibe codingu z min 100 pkt.
- highHN: Claude / Anthropic
HN posty z 'Claude' lub 'Anthropic' z min 100 pkt.
- highHugging Face Blog
Releases dla image, video, audio, 3D modeli. Część tech-heavy — Gemini relevance odfiltruje noise. Downgraded z critical: za duży volume na 'must-read' status.
- highIndyDevDan
Claude Code power user, prompty, hooki.
- highInterconnects (Nathan Lambert)
AI policy + research analysis. Niska hype rate, opinionated.
- highLatent Space
Podcast + blog Swyx — wywiady z founderami i deep dives engineeringowe.
- highMatt Wolfe
Comprehensive AI tools weekly digest. ~700K subs.
- highMatthew Berman
AI news, model release reviews, agent demos. Wysoki output.
- highr/aivideo
Community AI video — Sora, Veo, Runway, Kling, LTX. Co naprawdę zaskakuje twórców.
- highr/ClaudeAI
Społeczność Claude'a — power users, tipy, problemy.
- highr/LocalLLaMA
Open-source LLMs, lokalne uruchamianie, benchmarks bez hype.
- highr/StableDiffusion
Największa community open-source image gen (700k+ users). Premiery modeli, LoRA, ComfyUI workflows.
- highRiley Brown
Vibe coding, AI builder workflows, Cursor + Claude tutorials.
- highThe Decoder
Niemiecki AI news outlet po angielsku, dobre breaking news.
- highTheo - t3.gg
TypeScript + AI dev workflows. Hot takes, narrative-driven.
- highYannic Kilcher
Paper reviews i deep dives w research AI.
- lowAI Weirdness
Janelle Shane — playful AI experiments, image gen quirks. Niski volume, unikalna perspektywa.
- mediumbycloud
AI papers digestible — między 2MP a Yannic Kilcher.
- mediumCreative Bloq
Design industry — gdzie AI ingeruje w klasyczne dyscypliny graficzne.
- mediumFireship
100-sec format, often AI/LLM + tech news.
- mediumfxguide
VFX i film industry — coraz więcej AI w pipeline. Profesjonalna perspektywa.
- mediumGreg Isenberg
Solo founder vibe — buduje produkty z AI, podcasty z indie hackers.
- mediumr/ChatGPTCoding
Vibe coding tipy, IDE setupy, prompty. Mix wszystkich modeli.
- mediumr/comfyui
ComfyUI workflows — custom nodes, JSON workflows, optymalizacje.
- mediumr/midjourney
Midjourney community — premiery v7+, style references, prompt patterns.
- mediumr/runwayml
Runway-specific community — premiery features, prompt patterns, comparisons z konkurencją.
- mediumr/SunoAI
Suno music gen community — nowe wersje modelu, lyric prompting techniques. Audio AI ma slaby RSS ecosystem.
- mediumTina Huang
AI workflows for data science, practical applications.
- mediumTwo Minute Papers
Krótkie streszczenia papers AI, świetne dla szybkiego scan'a.
- mediumWes Roth
AI news z bardziej clickbaitowym tonem — filtr Gemini odsiewa hype.

I built a tool to mix two artists on one image with region masks — Van Gogh + Picasso, no training, arbitrary refs
Mix different artistic styles in specific parts of an image using masks and IP-Adapters without any training or fine-tuning.
A new open-source tool allows users to apply distinct artistic styles to specific regions of an image using spatial masks. Built on Stable Diffusion 1.5, the system utilizes ControlNet (Canny and Tile) for structural integrity and two IP-Adapters for style injection. The technical core involves spatial routing, where each adapter's contribution is masked within the cross-attention layers to prevent 'muddy' averaging of styles. It offers three modes: global mixing, painterly emphasis, and region-specific stylization. While effective, the author notes that aggressive style weights can distort realistic faces and small color details. The project includes a GitHub repository with a Colab notebook and a Hugging Face Space for testing.
r/StableDiffusion·tooling·05/07/2026, 09:24 AM·/u/Longjumping_Gur_937
LTX2.3 + ID LoRS + Prompt relay + Keyframes
Discover a powerful, all-in-one workflow for Stable Diffusion that simplifies creating AI videos with consistent characters, dynamic prompts, and advanced animation techniques.
A Reddit user, /u/Brief-Leg-8831, shared a comprehensive workflow on Civitai for generating advanced AI videos using Stable Diffusion. This 'all-in-one' setup integrates several powerful techniques including LTX2.3, ID LoRA for character consistency, Prompt relay for dynamic narrative progression, ControlNet for precise pose control, and Keyframes for animation timing. The workflow also incorporates a detailer, upscaler, and custom audio synchronization, offering a robust solution for creating complex and high-quality AI-generated video content. It addresses common challenges in AI video production by combining multiple tools into a streamlined process.
r/StableDiffusion·tooling·05/06/2026, 04:03 PM·/u/Brief-Leg-8831
I used Blender as a layout tool for AI video generation — here's the full workflow
Learn how to use Blender's 3D environment to gain precise spatial and camera control over AI video generation, solving common consistency issues.
The author demonstrates a hybrid workflow using Blender as a spatial layout tool to control AI video generation. By setting up basic 3D geometry and camera movements in Blender, they create a consistent structural reference that guides the AI's output. This method addresses the common issue of temporal and spatial instability found in pure text-to-video models. The workflow involves rendering simple 'graybox' scenes or depth maps from Blender and passing them through ControlNet or image-to-video pipelines like Stable Video Diffusion or Runway. It bridges the gap between precise 3D control and the aesthetic flexibility of generative AI, allowing for professional-grade shot composition and predictable movement.
r/aivideo·tutorial·05/05/2026, 04:19 PM·/u/waterarttrkglBadass professional workflow - How High-Effort AI Usage Looks
Learn how to combine manual 3D modeling and environment building with AI to create professional-grade animations with full control over consistency.
This workflow demonstrates a sophisticated bridge between traditional 3D artistry and AI generation. Instead of relying on text prompts alone, the creator uses manual robot modeling and 3D meshes to provide structural guidance for the AI. The process covers the entire pipeline from initial concept to environment creation and final frame generation. It highlights a shift towards high-effort AI usage where the model acts as a rendering engine for human-directed geometry. This approach solves common consistency issues in AI animation by grounding the output in 3D space, ensuring that movements and perspectives remain stable across frames.
r/StableDiffusion·tutorial·05/05/2026, 08:48 AM·/u/LazyChamberlain
LTX-2.3 + Union Control LoRA (8GB VRAM)
Generate high-quality 1280x640 LTX-2.3 videos with precise control on an 8GB VRAM GPU using this optimized ComfyUI workflow.
A new ComfyUI workflow demonstrates high-resolution video generation (1280x640) using the LTX-2.3 model on consumer-grade hardware with only 8GB of VRAM. By integrating the Union Control LoRA, users can achieve precise structural control over the video output, which was previously difficult on low-memory GPUs. The author provides a complete package including a Hugging Face repository for the workflow and a step-by-step YouTube tutorial. This release is significant for the creative community as it lowers the barrier to entry for high-quality AI cinematography. The pipeline uses Nano Banana for the initial frame generation before passing it to LTX-2.3 for temporal consistency.
r/comfyui·tooling·05/05/2026, 02:14 AM·/u/big-boss_97
Relevance auto-scored by LLM (0–10). List shows top 30 from the last 7 days.