AI pulse last 7 days
Daily AI pulse from YouTube, blogs, Reddit, HN. Ruthlessly filtered.
Sources (41)▶
- criticalAndrej Karpathy
Były dyrektor AI w Tesli, OpenAI cofounder. Każde video to gold.
- criticalAnthropic
Oficjalny kanał Anthropic. Każdy release Claude'a.
- criticalComfyUI Blog
Release log dla integracji ComfyUI — Luma Uni-1, GPT Image 2, ACE-Step music gen, Seedance. Pokrywa video+image+music+workflow.
- criticalOpenAI Blog
Oficjalny blog OpenAI. Wszystkie release.
- criticalSimon Willison's Weblog
Najlepszy 'thinker' AI. Codzienne posty, deep insights, niska hype rate.
- highAI Explained
Głęboka analiza papers i benchmarków, niska hype rate.
- highAI Jason
Praktyczne tutoriale Claude Code, MCP, workflow vibe codingu.
- highBen's Bites
Daily AI digest, creator-friendly tone. Codex, model releases, agentic AI.
- highCole Medin
Vibe coding + agentic workflows + Claude Code MCP integrations.
- highFal AI Blog
Fal hostuje większość nowych AI image/video modeli — ich blog to wczesne sygnały premier.
- highHN: 3D & Gaussian Splatting
HN signal dla 3D generative — Gaussian Splatting, NeRF, image-to-3D. Próg 20 bo niszowa kategoria (top historic 182pts).
- highHN: AI agents / MCP
HN posty o agentach, MCP, vibe codingu z min 100 pkt.
- highHN: Claude / Anthropic
HN posty z 'Claude' lub 'Anthropic' z min 100 pkt.
- highHugging Face Blog
Releases dla image, video, audio, 3D modeli. Część tech-heavy — Gemini relevance odfiltruje noise. Downgraded z critical: za duży volume na 'must-read' status.
- highIndyDevDan
Claude Code power user, prompty, hooki.
- highInterconnects (Nathan Lambert)
AI policy + research analysis. Niska hype rate, opinionated.
- highLatent Space
Podcast + blog Swyx — wywiady z founderami i deep dives engineeringowe.
- highMatt Wolfe
Comprehensive AI tools weekly digest. ~700K subs.
- highMatthew Berman
AI news, model release reviews, agent demos. Wysoki output.
- highr/aivideo
Community AI video — Sora, Veo, Runway, Kling, LTX. Co naprawdę zaskakuje twórców.
- highr/ClaudeAI
Społeczność Claude'a — power users, tipy, problemy.
- highr/LocalLLaMA
Open-source LLMs, lokalne uruchamianie, benchmarks bez hype.
- highr/StableDiffusion
Największa community open-source image gen (700k+ users). Premiery modeli, LoRA, ComfyUI workflows.
- highRiley Brown
Vibe coding, AI builder workflows, Cursor + Claude tutorials.
- highThe Decoder
Niemiecki AI news outlet po angielsku, dobre breaking news.
- highTheo - t3.gg
TypeScript + AI dev workflows. Hot takes, narrative-driven.
- highYannic Kilcher
Paper reviews i deep dives w research AI.
- lowAI Weirdness
Janelle Shane — playful AI experiments, image gen quirks. Niski volume, unikalna perspektywa.
- mediumbycloud
AI papers digestible — między 2MP a Yannic Kilcher.
- mediumCreative Bloq
Design industry — gdzie AI ingeruje w klasyczne dyscypliny graficzne.
- mediumFireship
100-sec format, often AI/LLM + tech news.
- mediumfxguide
VFX i film industry — coraz więcej AI w pipeline. Profesjonalna perspektywa.
- mediumGreg Isenberg
Solo founder vibe — buduje produkty z AI, podcasty z indie hackers.
- mediumr/ChatGPTCoding
Vibe coding tipy, IDE setupy, prompty. Mix wszystkich modeli.
- mediumr/comfyui
ComfyUI workflows — custom nodes, JSON workflows, optymalizacje.
- mediumr/midjourney
Midjourney community — premiery v7+, style references, prompt patterns.
- mediumr/runwayml
Runway-specific community — premiery features, prompt patterns, comparisons z konkurencją.
- mediumr/SunoAI
Suno music gen community — nowe wersje modelu, lyric prompting techniques. Audio AI ma slaby RSS ecosystem.
- mediumTina Huang
AI workflows for data science, practical applications.
- mediumTwo Minute Papers
Krótkie streszczenia papers AI, świetne dla szybkiego scan'a.
- mediumWes Roth
AI news z bardziej clickbaitowym tonem — filtr Gemini odsiewa hype.

I built a tool to mix two artists on one image with region masks — Van Gogh + Picasso, no training, arbitrary refs
Mix different artistic styles in specific parts of an image using masks and IP-Adapters without any training or fine-tuning.
A new open-source tool allows users to apply distinct artistic styles to specific regions of an image using spatial masks. Built on Stable Diffusion 1.5, the system utilizes ControlNet (Canny and Tile) for structural integrity and two IP-Adapters for style injection. The technical core involves spatial routing, where each adapter's contribution is masked within the cross-attention layers to prevent 'muddy' averaging of styles. It offers three modes: global mixing, painterly emphasis, and region-specific stylization. While effective, the author notes that aggressive style weights can distort realistic faces and small color details. The project includes a GitHub repository with a Colab notebook and a Hugging Face Space for testing.
r/StableDiffusion·tooling·05/07/2026, 09:24 AM·/u/Longjumping_Gur_937Is anyone actually getting good results with Flux2.DEV?
If you're struggling to get sharp, realistic images from Flux2.DEV, you're not alone; a user reports consistent issues with hazy outputs and a limited LoRA ecosystem, seeking comm…
A Reddit user on r/StableDiffusion, /u/Extension-Yard1918, has reported persistent issues achieving sharp, realistic images with the Flux2.DEV model over several months of testing. Despite efforts like increasing resolution and step count, and experimenting with different samplers and settings, the generated outputs consistently appear hazy, soft, or foggy, failing to match the quality of models like Z-Image Turbo. The user also notes a weak image editing feature and a nearly nonexistent LoRA ecosystem, questioning if the problem lies with the model's training data, VAE, scheduler, or their own workflow. They are seeking practical advice and specific settings from the community to unlock Flux2.DEV's potential.
r/StableDiffusion·opinion·05/07/2026, 09:15 AM·/u/Extension-Yard1918
Clippy Reloaded - a really sarky useful Clipboard node with no click.
Automatically import your system clipboard into ComfyUI workflows every time you queue a prompt, eliminating manual pasting.
Clippy Reloaded is a custom node for ComfyUI designed to streamline the process of getting text into your workflows. Instead of manually pasting text into a node, this tool automatically pulls whatever is currently in your system clipboard the moment you queue a prompt. This is particularly useful for users who frequently copy prompts, descriptions, or parameters from external websites or LLM chats. The node eliminates repetitive clicking and pasting, acting as a dynamic input source. It is available as an open-source repository on GitHub for easy integration into existing ComfyUI setups.
r/StableDiffusion·tooling·05/07/2026, 12:11 AM·/u/shootthesound
Ernie Image Lora training - my take
Practical insights and visual benchmarks for training LoRAs on the Ernie model, highlighting necessary adjustments to standard training workflows.
The author presents their findings and visual results from training a LoRA on the Ernie image model, a less common alternative to the Stable Diffusion ecosystem. The post includes specific technical insights into the training process, highlighting how hyperparameters like learning rate and rank need adjustment compared to standard SDXL workflows. Visual benchmarks provided via Imgur demonstrate the model's proficiency in handling complex architectural details and specific artistic styles. This contribution is particularly valuable for users looking to diversify their toolkit beyond mainstream models and understand the nuances of cross-architecture fine-tuning. It serves as both a technical guide and a proof-of-concept for the Ernie model's flexibility.
r/StableDiffusion·tutorial·05/06/2026, 10:53 PM·/u/malcolmrey
My Reference Latent Node including Auto Masking and Timesteps per image is out tomorrow
A new ComfyUI node that offers precise control over reference images through auto-masking and per-image timestep scheduling.
Developer /u/shootthesound has released ReferenceLatentPlus, a new custom node for ComfyUI designed to refine how reference images influence generations. The tool introduces auto-masking capabilities and allows users to set specific timesteps for each reference image, providing granular control over when and how much a source image affects the output. It includes integrated VAE input and maximum resolution controls, simplifying the pipeline for piping multiple images directly into a workflow. This release addresses the need for more precise element extraction from source material without complex manual masking. The node is now publicly available on GitHub for integration into existing Stable Diffusion setups.
r/StableDiffusion·tooling·05/06/2026, 10:31 PM·/u/shootthesoundIs there any interest for a Character dataset evaluation script ?
A new tool is being developed to help curate LoRA training datasets by detecting face mirroring and scoring image quality and variety.
A Reddit user has developed a Python script with a Gradio interface designed to optimize datasets for training LoRAs of real people. The tool addresses two specific problems: detecting mirrored faces to prevent unnaturally symmetrical results and providing a relevancy score based on image quality and variety. By filtering out redundant or low-quality images, the script aims to improve the final model's fidelity. While currently in the feedback stage, the author is gauging community interest before a public release. This could be a valuable utility for hobbyists struggling with manual dataset curation.
r/StableDiffusion·tooling·05/06/2026, 07:54 PM·/u/HumbleSousVideGeek
Anyone else tried this RefineAnything LoRA? Pretty impressed so far
A new ComfyUI plugin and LoRA workflow for surgical image refinement, perfect for fixing text, logos, and small details without affecting the rest of the image.
The RefineAnything project provides a specialized LoRA and workflow for surgical image repairs, specifically targeting text, logos, and product labels. A new ComfyUI plugin, ComfyUI-RefineNode, has been released to automate the manual labor of mask preparation, reference alignment, and pasting back the refined region. The plugin is model-agnostic, meaning it can enhance any local detail repair workflow, not just the RefineAnything LoRA. It supports both scribble masks and bounding boxes, ensuring the rest of the image remains 100% untouched. A technical tip from the developer suggests avoiding the 'index_timestep_zero' method to prevent noticeable color shifts during the process.
r/StableDiffusion·tooling·05/06/2026, 07:32 PM·/u/liangkun43
[Z-Image] REALSTAGRAM_ZIMG — subtle realism LoRA for Z-Image Turbo (works with any character LoRA)
Enhance Z-Image Turbo generations with a subtle, candid Instagram realism LoRA that stacks perfectly with character models.
REALSTAGRAM_ZIMG is a new realism-enhancing LoRA specifically designed for the Z-Image Turbo and De-Turbo models. It aims to shift image outputs away from the typical "AI-perfect" look toward a more amateur, candid Instagram aesthetic. The LoRA is lightweight (Rank 64, 325 MB) and does not require a trigger word, making it easy to integrate into existing prompts. It is optimized for stacking with character LoRAs at a strength of 0.2 to 0.6 to maintain character identity while adding subtle texture and lighting improvements. A ComfyUI workflow is provided to help users get started immediately.
r/StableDiffusion·tooling·05/06/2026, 06:35 PM·/u/Existing-House1230
LTX2.3 + ID LoRS + Prompt relay + Keyframes
Discover a powerful, all-in-one workflow for Stable Diffusion that simplifies creating AI videos with consistent characters, dynamic prompts, and advanced animation techniques.
A Reddit user, /u/Brief-Leg-8831, shared a comprehensive workflow on Civitai for generating advanced AI videos using Stable Diffusion. This 'all-in-one' setup integrates several powerful techniques including LTX2.3, ID LoRA for character consistency, Prompt relay for dynamic narrative progression, ControlNet for precise pose control, and Keyframes for animation timing. The workflow also incorporates a detailer, upscaler, and custom audio synchronization, offering a robust solution for creating complex and high-quality AI-generated video content. It addresses common challenges in AI video production by combining multiple tools into a streamlined process.
r/StableDiffusion·tooling·05/06/2026, 04:03 PM·/u/Brief-Leg-8831
Thanks to the sub my silly node and workflow got 3k downloads overnight, therefore I fixed some bugs, unified some features, and uploaded the latest and the greatest version to HF.
A new ComfyUI node that automates character consistency and scene composition using a structured Qwen-based procedural prompting system.
The ComfyUI Character Composer is a procedural prompt system designed to streamline character consistency and scene composition. Built upon the Qwen-Image-Edit-Rapid-AIO ecosystem, it provides a structured approach to generation, reducing the need for manual LLM prompting or copy-pasting. The tool features a unified txt2img and img2img workflow and utilizes a SFW JSON library for managing assets. Following a viral reception on Reddit with over 3,000 downloads, the developer has updated the node with bug fixes and unified features. It aims to offer more controllable generation for users working with complex character-driven workflows.
r/StableDiffusion·tooling·05/06/2026, 03:14 PM·/u/Mundane-Ad-5737
Testing VCI integrity with high-end commercial beauty textures in a low-key, high-saturation setup.
Achieve professional commercial beauty standards and structural consistency in low-light environments using a mobile-based AI workflow.
This project explores Visual Consistency Integrity (VCI) within a mobile-based Stable Diffusion workflow, focusing on high-end commercial beauty standards. The creator demonstrates how to achieve flawless skin textures that remain photorealistic even under challenging low-key, high-saturation lighting conditions. By comparing portraits with full-body shots featuring medieval armor, the experiment tests structural consistency and light reflection stability in low-light environments. The medieval armor acts as a complex subject to evaluate how the model handles intricate surfaces and deep shadows. This demonstration showcases the potential for professional-grade aesthetic results produced entirely on mobile devices.
r/StableDiffusion·creative_work·05/06/2026, 11:15 AM·/u/Victoria-rosemont
ComfyUI with co-founder Yannik Marek (ComfyAnonymous)
A deep dive with the creator of ComfyUI on how node-based AI workflows are moving from experimental hacks to professional VFX production standards.
This podcast episode features an interview with Yannik Marek, the creator of ComfyUI known as ComfyAnonymous, discussing the tool's journey from a personal experiment to a professional industry standard. They explore how the node-based architecture allows for precise control over Stable Diffusion pipelines, making it indispensable for high-end VFX work. The discussion covers the transition to Comfy Org and the focus on stability and performance for enterprise environments. Marek explains the rationale behind the modular design, which enables rapid integration of new models and techniques. This is a deep dive into the technical philosophy that has made ComfyUI the preferred interface for advanced AI creators.
fxguide·tooling·05/06/2026, 09:38 AM·Mike Seymour
GTA 70s - Teaser Trailer (Alternative Version): Z-image Turbo - Flux Klein 9b - Wan 2.2
A high-quality 70s-style GTA trailer showcase using Flux and Wan 2.2, complete with downloadable ComfyUI workflows for replication.
This project showcases a fan-made 'GTA 70s' teaser trailer created using a sophisticated AI video pipeline. The creator utilized Flux Klein 9b for high-quality image generation and Wan 2.2 for video synthesis, achieving a distinct 70s cinematic aesthetic. Unlike many AI-generated videos that rely on heavy filters, this version focuses on clean film colors and realistic motion. Crucially, the author shared the full ComfyUI workflows via Google Drive, allowing the community to study and replicate the specific generation techniques. It serves as a practical benchmark for what is currently achievable with open-weight video models and fine-tuned Flux variants.
r/StableDiffusion·creative_work·05/06/2026, 08:36 AM·/u/MayaProphecyInstall Stable Diffusion WebUI Forge easily on Windows: portable one-click installer for Forge Classic + Forge Neo
Install and switch between Stable Diffusion Forge Classic and Neo on Windows easily with a single portable one-click batch script.
A new community-developed portable batch script simplifies the installation of Stable Diffusion WebUI Forge on Windows. This tool automates the entire setup process, including Python, Git, and virtual environment configuration, which are often major hurdles for non-technical users. It offers a choice between Forge Classic for stability and Forge Neo for experimental features. The installation is self-contained, meaning it won't clutter your system and can be easily moved or deleted. This utility is particularly useful for those looking for a high-performance alternative to the standard Automatic1111 interface without the manual setup headache.
r/StableDiffusion·tooling·05/05/2026, 11:52 PM·/u/Merserk13
Wireframe - Flux.2 Klein 9b style LORA
New 'Wireframe' style LoRA for Flux.2 Klein 9b enables technical, mesh-like aesthetics in AI generations using the trigger word 'dvr_wf_style'.
Developer Dever has released a specialized Wireframe style LoRA designed specifically for the Flux.2 Klein 9b distilled model. This LoRA allows users to generate or edit images to have a technical, 3D-mesh aesthetic using the trigger word 'dvr_wf_style'. It was trained on the 9b base as a text-to-image model but demonstrates high flexibility in image-to-image editing tasks. The weights are hosted on Huggingface, where the author maintains a repository of various style LoRAs for the Flux ecosystem. This release is particularly relevant for creators looking for architectural or blueprint-like visuals within the Flux.2 framework.
r/StableDiffusion·model_release·05/05/2026, 10:20 PM·/u/TheDudeWithThePlanI built a dual-monitor image curator for sorting large Stable Diffusion output folders (looking for feedback)
A new local desktop tool designed specifically for fast sorting and tagging of large Stable Diffusion image folders using dual monitors.
A developer has released a specialized desktop tool aimed at solving the image bloat problem for Stable Diffusion users. The application allows for rapid curation of large output folders by tagging images as favorites or junk and jumping through batches of up to 100 files. A standout feature is the dual-monitor support, which keeps one screen dedicated to clean image viewing while the other handles controls. The tool operates entirely locally, ensuring privacy and speed compared to cloud-based solutions or generic file explorers. It addresses a specific workflow bottleneck where standard OS tools fail to handle high-volume AI generation efficiently.
r/StableDiffusion·tooling·05/05/2026, 08:46 PM·/u/ZAI_DevLocal Dream 2.4.3 - SDXL support, tag autocomplete and more
You can now run SDXL and PonyXL models locally on high-end Snapdragon devices using NPU acceleration with Local Dream 2.4.3.
Local Dream version 2.4.3 introduces significant updates for mobile generative AI, specifically targeting high-end Snapdragon hardware. The app now supports SDXL, Illustrious, and PonyXL models by leveraging the NPU on Snapdragon 8 Gen 3 and Elite chips. New quality-of-life features include a tag autocomplete system based on CSV imports and a real-time token counter for prompt engineering. Additionally, the inclusion of the LCM scheduler allows for faster image generation with fewer steps. This update bridges the gap between desktop and mobile local generation for power users who want to run heavy models on the go.
r/StableDiffusion·tooling·05/05/2026, 11:53 AM·/u/mikemend
Converting 2D animations to 3D with LTX 2.3 Lora
Transform 2D animations into depth-rich 3D videos using LTX-Video 2.3 and a specific LoRA workflow for improved spatial consistency.
This workflow demonstrates a method for converting flat 2D animations into 3D-style videos using the LTX-Video 2.3 model and a specialized LoRA. By leveraging the temporal consistency of the LTX architecture, the technique goes beyond simple depth effects to create genuine spatial volume and realistic lighting. The process involves using existing 2D footage as a structural reference while the LoRA guides the model to reinterpret the scene with 3D depth. This provides a powerful tool for creators to modernize 2D assets or generate complex parallax movements without traditional 3D software. It highlights the growing ecosystem of fine-tuned adapters for open-source video generation models.
r/StableDiffusion·tutorial·05/05/2026, 09:09 AM·/u/CQDSNBadass professional workflow - How High-Effort AI Usage Looks
Learn how to combine manual 3D modeling and environment building with AI to create professional-grade animations with full control over consistency.
This workflow demonstrates a sophisticated bridge between traditional 3D artistry and AI generation. Instead of relying on text prompts alone, the creator uses manual robot modeling and 3D meshes to provide structural guidance for the AI. The process covers the entire pipeline from initial concept to environment creation and final frame generation. It highlights a shift towards high-effort AI usage where the model acts as a rendering engine for human-directed geometry. This approach solves common consistency issues in AI animation by grounding the output in 3D space, ensuring that movements and perspectives remain stable across frames.
r/StableDiffusion·tutorial·05/05/2026, 08:48 AM·/u/LazyChamberlain
A new open weights image model appears in ArtificialAnalysis. Outperforming Flux.2 Pro and Z Image Turbo.
A new open-weights image model has topped the ArtificialAnalysis leaderboard, outperforming Flux.2 Pro and Z Image Turbo in human preference tests.
A new open-weights image generation model has surfaced on the ArtificialAnalysis leaderboard, claiming the top spot over established models like Flux.2 Pro and Z Image Turbo. This model's performance in Elo-based human preference rankings suggests a significant leap in quality for the open-source community. This development is crucial as it challenges the dominance of closed-source or 'pro' tier models in visual fidelity and prompt adherence. The community is currently dissecting the model's architecture and availability for local deployment. Early data indicates superior handling of complex textures and lighting compared to its predecessors, marking a potential shift in the state-of-the-art for local image generation.
r/StableDiffusion·model_release·05/04/2026, 07:07 PM·/u/Murky_Foundation5528
Relevance auto-scored by LLM (0–10). List shows top 30 from the last 7 days.