AI pulse last 7 days
Daily AI pulse from YouTube, blogs, Reddit, HN. Ruthlessly filtered.
Sources (41)▶
- criticalAndrej Karpathy
Były dyrektor AI w Tesli, OpenAI cofounder. Każde video to gold.
- criticalAnthropic
Oficjalny kanał Anthropic. Każdy release Claude'a.
- criticalComfyUI Blog
Release log dla integracji ComfyUI — Luma Uni-1, GPT Image 2, ACE-Step music gen, Seedance. Pokrywa video+image+music+workflow.
- criticalOpenAI Blog
Oficjalny blog OpenAI. Wszystkie release.
- criticalSimon Willison's Weblog
Najlepszy 'thinker' AI. Codzienne posty, deep insights, niska hype rate.
- highAI Explained
Głęboka analiza papers i benchmarków, niska hype rate.
- highAI Jason
Praktyczne tutoriale Claude Code, MCP, workflow vibe codingu.
- highBen's Bites
Daily AI digest, creator-friendly tone. Codex, model releases, agentic AI.
- highCole Medin
Vibe coding + agentic workflows + Claude Code MCP integrations.
- highFal AI Blog
Fal hostuje większość nowych AI image/video modeli — ich blog to wczesne sygnały premier.
- highHN: 3D & Gaussian Splatting
HN signal dla 3D generative — Gaussian Splatting, NeRF, image-to-3D. Próg 20 bo niszowa kategoria (top historic 182pts).
- highHN: AI agents / MCP
HN posty o agentach, MCP, vibe codingu z min 100 pkt.
- highHN: Claude / Anthropic
HN posty z 'Claude' lub 'Anthropic' z min 100 pkt.
- highHugging Face Blog
Releases dla image, video, audio, 3D modeli. Część tech-heavy — Gemini relevance odfiltruje noise. Downgraded z critical: za duży volume na 'must-read' status.
- highIndyDevDan
Claude Code power user, prompty, hooki.
- highInterconnects (Nathan Lambert)
AI policy + research analysis. Niska hype rate, opinionated.
- highLatent Space
Podcast + blog Swyx — wywiady z founderami i deep dives engineeringowe.
- highMatt Wolfe
Comprehensive AI tools weekly digest. ~700K subs.
- highMatthew Berman
AI news, model release reviews, agent demos. Wysoki output.
- highr/aivideo
Community AI video — Sora, Veo, Runway, Kling, LTX. Co naprawdę zaskakuje twórców.
- highr/ClaudeAI
Społeczność Claude'a — power users, tipy, problemy.
- highr/LocalLLaMA
Open-source LLMs, lokalne uruchamianie, benchmarks bez hype.
- highr/StableDiffusion
Największa community open-source image gen (700k+ users). Premiery modeli, LoRA, ComfyUI workflows.
- highRiley Brown
Vibe coding, AI builder workflows, Cursor + Claude tutorials.
- highThe Decoder
Niemiecki AI news outlet po angielsku, dobre breaking news.
- highTheo - t3.gg
TypeScript + AI dev workflows. Hot takes, narrative-driven.
- highYannic Kilcher
Paper reviews i deep dives w research AI.
- lowAI Weirdness
Janelle Shane — playful AI experiments, image gen quirks. Niski volume, unikalna perspektywa.
- mediumbycloud
AI papers digestible — między 2MP a Yannic Kilcher.
- mediumCreative Bloq
Design industry — gdzie AI ingeruje w klasyczne dyscypliny graficzne.
- mediumFireship
100-sec format, often AI/LLM + tech news.
- mediumfxguide
VFX i film industry — coraz więcej AI w pipeline. Profesjonalna perspektywa.
- mediumGreg Isenberg
Solo founder vibe — buduje produkty z AI, podcasty z indie hackers.
- mediumr/ChatGPTCoding
Vibe coding tipy, IDE setupy, prompty. Mix wszystkich modeli.
- mediumr/comfyui
ComfyUI workflows — custom nodes, JSON workflows, optymalizacje.
- mediumr/midjourney
Midjourney community — premiery v7+, style references, prompt patterns.
- mediumr/runwayml
Runway-specific community — premiery features, prompt patterns, comparisons z konkurencją.
- mediumr/SunoAI
Suno music gen community — nowe wersje modelu, lyric prompting techniques. Audio AI ma slaby RSS ecosystem.
- mediumTina Huang
AI workflows for data science, practical applications.
- mediumTwo Minute Papers
Krótkie streszczenia papers AI, świetne dla szybkiego scan'a.
- mediumWes Roth
AI news z bardziej clickbaitowym tonem — filtr Gemini odsiewa hype.

Roblox Scientoloty Speedrun made with SuperGrok
See a humorous AI-generated video "speedrun" in a Roblox style, showcasing the creative capabilities of the SuperGrok tool for generating unique content.
A Reddit user, /u/ginadaspokemon, shared a unique AI-generated video titled "Roblox Scientoloty Speedrun" created with a tool called SuperGrok. This creative work showcases the potential of AI video generation to produce highly specific and humorous content. The video adopts a distinct Roblox-like aesthetic, demonstrating SuperGrok's capability to generate stylized narratives. It provides a concrete example of how AI tools can be leveraged by hobbyists and creative non-developers to create engaging and niche video content, moving beyond generic outputs. This highlights the evolving landscape of AI-powered creative expression in video.
r/aivideo·creative_work·05/07/2026, 12:36 PM·/u/ginadaspokemon
Made this with Nano + Kling 3
See a user-generated AI video created with Nano and Kling 3 to get a sense of current creative capabilities and tool combinations in AI video generation.
A Reddit user, /u/Entire-Turnover-8560, posted an AI-generated video created using a combination of tools identified as "Nano" and "Kling 3". This submission on r/aivideo serves as a practical demonstration of current AI video generation capabilities, particularly for creative hobbyists interested in the output quality and stylistic potential of these models. While specific details about "Nano" are not provided, "Kling 3" likely refers to Kuaishou's advanced video generation model, known for its high-fidelity outputs. The post highlights how these tools can be combined to produce compelling visual content, offering inspiration for those exploring AI in creative workflows.
r/aivideo·creative_work·05/07/2026, 11:25 AM·/u/Entire-Turnover-8560
So Far This is My Favorite Use-Case for LTX 2.3/ComfyUI
Discover a practical workflow for using the LTX 2.3 video model in ComfyUI to achieve high-quality, consistent video generation on local hardware.
The Reddit community is exploring the capabilities of LTX 2.3, a new video generation model, specifically within the ComfyUI node-based interface. This post demonstrates a high-quality use-case that highlights the model's strengths in temporal consistency and motion fidelity. LTX 2.3 is designed to be more accessible for local execution on consumer GPUs than previous state-of-the-art video models. The author's workflow provides a practical example of how to integrate this model into complex creative pipelines. This demonstration is particularly valuable for creators looking for alternatives to closed-source video tools like Runway or Luma.
r/StableDiffusion·tooling·05/07/2026, 08:33 AM·/u/optimisoprimeo
testing LTX 2.3 1.1 distilled on my gpu. pretty much decent for creating ugc content or short tiktok vlog.
Distilled LTX 2.3 enables fast, high-quality local video generation on mid-range GPUs like the RTX 4060 Ti when paired with the latest CUDA/Torch updates.
A user on r/comfyui demonstrates the performance of the distilled LTX 2.3 1.1 model for generating short-form video content locally. The test highlights significant performance gains when using updated software stacks, specifically Torch 2.11.0 and CUDA 13.0. Running on consumer-grade hardware (RTX 4060 Ti 16GB), the model is capable of producing decent quality UGC and TikTok-style vlogs. The post includes a link to the specific ComfyUI workflow used for these results. This release represents a step forward in making high-quality video generation accessible on mid-range local GPUs.
r/comfyui·tooling·05/07/2026, 08:10 AM·/u/aziib
testing LTX 2.3 v1.1 distilled on my gpu. pretty decent for creating ugc content or short tiktok vlog.
LTX 2.3 v1.1 distilled runs efficiently on mid-range consumer GPUs (RTX 4060 Ti) for short video content when using updated Torch and CUDA drivers.
A user report demonstrates the performance of LTX 2.3 v1.1 distilled for creating short-form video content like TikTok vlogs. Running on an RTX 4060 Ti 16GB, the model shows significant speed improvements when paired with PyTorch 2.11.0 and CUDA 13.0 in ComfyUI. The distilled version of the model is specifically optimized for faster inference while maintaining enough quality for social media use cases. The post highlights the importance of driver and library updates for maximizing performance on consumer-grade hardware, making high-quality video generation more accessible.
r/StableDiffusion·tooling·05/07/2026, 08:10 AM·/u/aziibKijai LTX 2.3 WIth 12 GB of VRam demo reel
You can now run the high-quality LTX 2.3 22B video model on a standard 12GB VRAM GPU using GGUF quantization and specialized ComfyUI workflows.
A user demonstrated that the LTX 2.3 22B video generation model can produce high-quality 8-second clips on consumer-grade hardware. By utilizing GGUF quantization and specific ComfyUI workflows developed by Kijai, the model fits within 12GB of VRAM, specifically tested on an RTX 3060 with 32GB of system RAM. This is a significant milestone as it brings state-of-the-art open-weight video generation to hobbyist setups. The shared resources include the GGUF model files and optimized workflows available on Civitai. This setup balances performance and accessibility, making long-form AI video generation more feasible for local execution without requiring enterprise-grade hardware.
r/comfyui·tooling·05/06/2026, 09:09 PM·/u/OfficeMagic1
Interactive Video Generation (Causal Forcing) - High Speed!
Generate high-speed interactive videos even on mid-range GPUs like the RTX 3060, with potential for real-time performance on high-end hardware.
Causal Forcing is a new approach to interactive video generation that emphasizes speed and efficiency. The release includes open-source code and models, with a community-repackaged version for ComfyUI. Performance benchmarks show that an RTX 3060 can generate a 2-second video (848x480) in just 11 seconds using only 4 steps. On high-end GPUs like the RTX 4090 or 5090, users report near real-time generation speeds. The model is lightweight, peaking at 6GB VRAM, making it accessible for hobbyists with mid-range hardware. This represents a significant step toward fluid, interactive AI video tools.
r/StableDiffusion·model_release·05/06/2026, 05:53 PM·/u/ZerOne82
Ella - [AI orchestrated music video generation | more info in comments]
Discover how AI can orchestrate and generate entire music videos, offering a new avenue for creative expression and automated visual storytelling synchronized with audio.
User /u/TasTepeler showcased "Ella," an AI-orchestrated music video generation project on r/aivideo. This initiative demonstrates a sophisticated approach to creating music videos where artificial intelligence manages the synchronization and visual composition in response to audio. The project highlights the growing capability of AI to move beyond simple image or video generation towards more complex, integrated creative tasks. It represents a significant step in automating the labor-intensive process of music video production, offering creative non-developers and hobbyists a glimpse into future possibilities for dynamic visual content creation.
r/aivideo·creative_work·05/06/2026, 04:01 PM·/u/TasTepeler
LTX2.3 + Prompt relay + Keyframes | 2027 ChatGPT self awareness event 😝
Master complex video transitions in ComfyUI using a comprehensive LTX2.3 workflow that integrates prompt relaying and keyframe control.
A new advanced ComfyUI workflow for the LTX2.3 video model has been shared, focusing on the synergy between prompt relaying and keyframes. The setup allows for complex narrative transitions and visual consistency by chaining prompts and managing motion via keyframes. Beyond basic generation, the workflow integrates ID LoRA for character consistency, ControlNet for structural guidance, and a detailer/upscaler pass for high-quality output. It also includes support for custom audio synchronization. While the author notes that the results can be finicky, the provided Civitai link offers a complete all-in-one solution for creators looking to push the boundaries of AI video.
r/comfyui·tooling·05/06/2026, 03:57 PM·/u/Brief-Leg-8831
GTA 70s - Teaser Trailer (Alternative Version): Z-image Turbo - Flux Klein 9b - Wan 2.2
A high-quality fan trailer demonstrating the synergy between Flux Klein 9b and Wan 2.2 for consistent, cinematic AI video generation.
This creative project showcases a 1970s-themed Grand Theft Auto teaser trailer created using a sophisticated AI pipeline in ComfyUI. The creator utilized Flux Klein 9b for image generation and Wan 2.2 for video synthesis, achieving a distinct vintage aesthetic. The workflow also incorporates Z-image Turbo, likely for rapid prototyping or specific style transfers. This piece serves as a benchmark for how hobbyists can combine multiple specialized models to produce high-fidelity, thematic video content. It highlights the rapid evolution of open-source video tools and their ability to maintain stylistic consistency across scenes.
r/comfyui·creative_work·05/06/2026, 08:36 AM·/u/MayaProphecy
Tencent is about to release an anime video model (AniMatrix).
Tencent is set to release AniMatrix, a specialized anime video generation model with open weights and inference code.
Tencent has announced the upcoming release of AniMatrix, a specialized video generation model focused on high-quality anime content. According to the accompanying ArXiv paper, the researchers intend to publicly release both the model weights and the inference code, a significant move in a field dominated by closed-source models. The project aims to solve common issues in AI animation, such as temporal consistency and stylistic accuracy specific to Japanese-style animation. By providing open access, Tencent is positioning itself as a major contributor to the open-source creative AI community. This release could provide a powerful new tool for hobbyists and professional animators who require more control than current proprietary web-based generators offer.
r/StableDiffusion·model_release·05/06/2026, 03:44 AM·/u/Total-Resort-3120
LTX2.3 8GB VRAM WorkFlow
Run the latest LTX2.3 video generation model on consumer-grade 8GB VRAM GPUs using this optimized ComfyUI workflow.
This Reddit post provides a specialized ComfyUI workflow designed to run the LTX2.3 video generation model on hardware with only 8GB of VRAM. LTX-Video is a high-quality open-weights model known for strong temporal consistency, but it typically demands significant GPU resources. By utilizing optimizations like model offloading or specific node configurations, this workflow makes high-end video generation accessible to users with mid-range consumer GPUs like the RTX 3060 or 4060. This is a practical solution for hobbyists who previously could not run the full model locally due to memory constraints.
r/comfyui·tooling·05/05/2026, 10:27 PM·/u/Extension-Yard1918Trying to use V2V to extend videos and create long-form in LTX2.3. Quality degrading over time.
Extending videos in LTX-2.3 using V2V workflows often leads to quality degradation after 30 seconds due to recursive referencing and artifact accumulation.
A user on r/comfyui is reporting significant quality loss when attempting to extend 10-second clips into 1-minute videos using the LTX-2.3 model. The process involves using Rune's V2V (Video-to-Video) workflow, which relies on the final 3 seconds of a previous segment to generate the next. By the 30-second mark, which is the third iteration, the visual fidelity begins to break down. This highlights a common 'drift' issue in recursive video generation where artifacts and noise accumulate over time. The discussion points to the limitations of current LTX-2.3 workflows for long-form content without more robust context management or latent refreshing.
r/comfyui·tooling·05/05/2026, 08:10 PM·/u/BarelyAI
GTA 70s - Teaser Trailer: Z-Image Turbo - Flux Klein 9b - Wan 2.2
A cinematic 'GTA 70s' fan trailer that includes downloadable workflows for Flux Klein 9b and Wan 2.2, optimized for 16GB VRAM hardware.
User MayaProphecy shared a high-quality teaser trailer for a conceptual 'GTA 70s' project, showcasing the synergy between several new generative models. The production utilizes Flux Klein 9b for base imagery and Wan 2.2 for video synthesis, achieving a consistent retro-cinematic aesthetic. Crucially, the author provided the full generation workflows via Google Drive, allowing hobbyists to study the node structures. The project was rendered on an RTX 5060 Ti 16GB, demonstrating that sophisticated video pipelines are now viable on mid-range consumer hardware. This serves as a benchmark for combining 'Turbo' speed models with high-fidelity video refiners.
r/StableDiffusion·creative_work·05/05/2026, 02:07 PM·/u/MayaProphecy
LTX2.3 8GB VRAM WorkFlow
Run the LTX2.3 video model on budget GPUs (8GB VRAM) using this optimized, multi-step ComfyUI workflow.
This Reddit post introduces a specialized ComfyUI workflow designed to run the LTX2.3 video generation model on GPUs with only 8GB of VRAM, such as the RTX 3060 Ti. Traditionally, high-end video models require significant hardware resources, but this optimization makes the technology accessible to hobbyists. The workflow achieves stability by generating initial video at a lower resolution at 24fps, then handling upscaling and frame interpolation as separate, decoupled steps. It supports both Text-to-Video and Image-to-Video modes, with the latter recommended for maintaining character consistency. This release provides a practical starting point for creative users who want to experiment with state-of-the-art video AI without expensive hardware upgrades.
r/StableDiffusion·tooling·05/05/2026, 12:46 PM·/u/Extension-Yard1918
My LTX 2.3 LoRA Training Journey: Fighting for VRAM even with a 5090
Training LTX 2.3 LoRAs on 32GB VRAM is viable by disabling audio modules and using official scripts, with results generalizing well to high-res video.
A detailed technical report on training a LoRA for the LTX 2.3 video model using an RTX 5090. The author highlights that AI-Toolkit proved unstable, leading them to use official training scripts refined with the help of Claude. To fit the training within 32GB of VRAM, it was mandatory to disable the audio module and limit resolution to 512x512 at 49 frames. Performance metrics showed 0.58 steps per second, with 1500 steps completed in 40 minutes. The resulting LoRA successfully captured specific 2D animation motion patterns and generalized well to higher resolutions and 121-frame sequences during inference.
r/StableDiffusion·tutorial·05/05/2026, 10:22 AM·/u/ovpresentme
Y'all might want to try this
New Causal-Forcing technique brings KV Cache and potential real-time frame generation to Wan models in ComfyUI.
The Causal-Forcing technique from Thu-ML is being integrated into ComfyUI via a new Pull Request, specifically targeting the Wan model architecture. This method allows for generating video frames sequentially with the benefit of KV Cache, which significantly optimizes memory and compute during inference. While the original researchers claim real-time performance on an RTX 4090, specific resolution details remain unconfirmed. The implementation in ComfyUI's core signifies a shift towards more efficient autoregressive video generation. This update is crucial for users looking to experiment with long-form video or interactive AI generation.
r/StableDiffusion·tooling·05/05/2026, 06:13 AM·/u/Altruistic_Heat_9531
LTX-2.3 + Union Control LoRA (8GB VRAM)
Generate high-quality 1280x640 LTX-2.3 videos with precise control on an 8GB VRAM GPU using this optimized ComfyUI workflow.
A new ComfyUI workflow demonstrates high-resolution video generation (1280x640) using the LTX-2.3 model on consumer-grade hardware with only 8GB of VRAM. By integrating the Union Control LoRA, users can achieve precise structural control over the video output, which was previously difficult on low-memory GPUs. The author provides a complete package including a Hugging Face repository for the workflow and a step-by-step YouTube tutorial. This release is significant for the creative community as it lowers the barrier to entry for high-quality AI cinematography. The pipeline uses Nano Banana for the initial frame generation before passing it to LTX-2.3 for temporal consistency.
r/comfyui·tooling·05/05/2026, 02:14 AM·/u/big-boss_97
LTX 2.3 Prompt Relay - Really good for concistency
Use the 'Prompt Relay' technique in ComfyUI to fix character flickering and maintain visual consistency in LTX 2.3 video generations.
A new workflow technique for LTX 2.3 called 'Prompt Relay' has been demonstrated to significantly improve character and environment consistency in generated videos. The method involves passing prompt information across frames or segments in a specific ComfyUI node setup to maintain visual coherence. This approach addresses the common issue of flickering or character morphing that plagues many open-source video models. By chaining prompt context, users can achieve more stable long-form or multi-shot sequences without losing the original artistic intent. The community is highlighting this as a practical solution for creators using LTX-Video checkpoints who need professional-grade stability.
r/comfyui·tooling·05/04/2026, 09:38 PM·/u/smereces
THE GAINS — Official Trailer (Sci-Fi Psychological Thriller)
A professional-grade sci-fi trailer showcasing how Runway's tools can be used for high-end narrative storytelling and cinematic consistency.
"THE GAINS" is a high-concept sci-fi psychological thriller trailer created by reddithaggz using Runway's generative video tools. Produced for the #RunwayBigPitchContest, the work demonstrates the current state of AI-driven cinematography, focusing on character consistency and atmospheric world-building. The trailer showcases sophisticated lighting and scene transitions that move beyond simple prompt-to-video outputs, suggesting a more directed approach to AI filmmaking. It serves as a practical example of how independent creators are leveraging Gen-3 Alpha or similar models to pitch complex narrative ideas that would traditionally require significant production budgets.
r/runwayml·creative_work·05/04/2026, 08:19 PM·/u/reddithaggz
April Wrapped
ComfyUI adds massive video (Seedance 2.0, Wan 2.7), music (Ace Step 1.5 XL), and SVG (Quiver) support, plus parallel API execution for speed.
ComfyUI's April update introduces a wide array of new models and features, significantly expanding its creative reach. Key additions include Seedance 2.0 and Wan 2.7 for advanced video generation, and Quiver for structured SVG (vector) output. Music generation gets a boost with Ace Step 1.5 XL and Sonilo's video-to-audio capabilities. On the technical side, the introduction of Parallel Job Execution via API allows for simultaneous workflow processing, offering a major productivity gain for production environments. The ComfyHub repository has also grown to nearly 500 community-shared workflows, making it easier to find pre-built solutions.
ComfyUI Blog·tooling·05/04/2026, 04:37 PM·Team at Comfy
Runway bringing ideas to life!
A high-quality example of achieving consistent claymation and stop-motion styles in Runway, proving AI can mimic complex analog animation textures.
This creative showcase demonstrates the use of Runway to achieve a consistent stop-motion and claymation aesthetic. The creator, Jesse's Devlog, highlights the difficulty of prompt engineering to maintain visual consistency and tactile textures across different shots. The resulting video mimics the jittery, handmade feel of traditional clay animation, which is notoriously difficult to replicate in AI video without significant flickering. While specific technical settings weren't disclosed, the project serves as a high-quality example of stylized video generation. It underscores the potential for AI to replace labor-intensive analog techniques for hobbyists and small studios.
r/runwayml·creative_work·05/01/2026, 02:23 AM·/u/JessesDevlog
Relevance auto-scored by LLM (0–10). List shows top 30 from the last 7 days.