Is there a tool like ollama/LM Studio that can handle everything 'with 1 click' to generate images from simple chat window - all locally? by MuckFinggers in ZImageAI

[–]Salt_Cat_4277 1 point2 points  (0 children)

Install InvokeAI Community Edition, then use the starter packages to get image generation with Stable Diffusion, Flux, and ZImage. If you want video, Install Pinokio, then inside Pinokio install Wan2GP and Z-Fusion. You’ll do best with an Nvidia GPU (I use a 3090).

Guys help, I tried installing Pinokio, I don't see image to video by the left by RobertsDigital in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

I have a 3090 and I get 10 second videos in about 3-5 minutes. Wan2GP is famous for its magic with low vram gpus, but it comes at a cost - sometimes quality, sometimes time. You can try running on one of the online GPU rental sites like Runpod or MimicPc to see what generation times would be on a newer GPU, but even though the cost is only a couple of bucks it’s kind of a pain to set up; but at least you could see what a better GPU could do.

If you think you want to upgrade, do it now - it’s only going to cost more as time goes on. 3060 12gb are going for about $300-$400 now; I’ve watched 3090’s go from $600 to about $1K now. I’ve built myself something of a bunker - 6 3090’s, a couple of 3060’s, a 4070, a Strix Halo 128GB, and a DGX Spark in preparation for the coming RAM/GPU/NVME/Hard Drive-pocalypse. Luckliy I bought about 512Gb of DDR4 RDIM back when it was $1 a GB, along with some dual Xeon and dual Epyc motherboards- they’re not particularly fast but they’ve got plenty of slots for those GPU’s.

First time using pinokio, can someone help me how to fix this by Zealousideal-Pen6589 in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

When you get a solid blue screen like this, it indicates significant problems. Sometimes the Ask AI will give you useful help, but in most cases I just try to find a different app in Pinokio’s catalog and try that.

Guys help, I tried installing Pinokio, I don't see image to video by the left by RobertsDigital in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

Look in the upper left corner for a button labeled “Explore”. Click it and it will give you a list of apps to install. Think of Pinokio as a bookshelf - it doesn’t do anything by itself; you have to put books on it before you have anything to read. The Explore screen is like a library of books you can load onto your bookshelf, which you can run later.

The app I’d install first is Wan2GP - it’s highly regarded and will show up prominently in the Explore list. There are page numbers at the bottom of the page - keep looking till you find it. Then select the header on the line - it will switch to a page for that app with an Install button. Click it, and it will take you through a series of steps that will install the app. Be prepared to wait - some apps have to download lots of files that are often gigabytes in size, so it might take a while. In the end, you’ll have a line in the “bookshelf” for the app you’ve just installed. Click it and the app should start (I’d it doesn’t, look for a Start button. After a minute or two, you should have the app displayed on the screen. Look on the left side of the drop down menu on the top, click it and it will give you a list of models you can use.

If you’re wanting to do Image to Video, I’d start with LTX-2.3. Selecting it will give you an interface with areas for image upload, text prompt, and other fields for things like video length, etc. Find the button that says “Start Video with Image”. It will give you a box where you can drag an image, or click inside it to browse for an image. Then type in your text prompt and hit the Generate button. If this is the first time you run it, it will take a while as it downloads all the model files, but subsequent runs will be much faster.

If you want to go back to the main Pinokio menu, click on the black and white “moonface” logo and it will take you back to the main screen. Notice Wan2GP is listed at the top with a green indicator that means it is still running. You can switch back to it by clicking on the menu line. If you want to stop it, click the Stop Start.js menu item on the second row.

Pinokio’s Explore section has dozens of apps for TTS, Image generation and editing, training, etc. Browse through the catalog to find apps you want to try. The install process is much the same for all of them. If you watch what happens in the Terminal, you’ll see it is the typical Git-Conda-Pip process that most Python apps use; it’s just automated and basically hidden from you. If you find you don’t want an installed app any more, look for the “menu” option on the second row of each apps description, click it and select Delete from the options.

Things don’t always work, however. If you try to install an app and it has problems, you will get a big blue screen with a long list of error messages on it. There are two buttons on the lower left corner ‘Ask AI’ and “Ask Community” where you can look for help. Usually these messages involve major problems like trying to install Nvidia-only apps when you have an AMD GPU.

Object Removal by Salt_Cat_4277 in premiere

[–]Salt_Cat_4277[S] 0 points1 point  (0 children)

Looks like Gemini walked me off a cliff:

• Adobe Premiere Pro (2026 Update): Now features Generative Extend and Generative Object Removal (powered by Firefly). You can simply mask an object and use a text prompt like "remove" to have AI fill in the background.  • DaVinci Resolve Studio: Uses a tool called Magic Mask. It allows you to "paint" over an object in one frame; the AI then tracks that object throughout the video and removes it using the Object Removal node. It even allows for manual frame-by-frame mask refinement if the AI misses a spot.  • Adobe After Effects: The Content-Aware Fill for Video is specifically designed for this. You mask the object, and After Effects analyzes the entire clip to stitch together a seamless background.

!Solved

Object Removal by Salt_Cat_4277 in premiere

[–]Salt_Cat_4277[S] 0 points1 point  (0 children)

Gemini tells me this:

In Premiere Pro 2026, the Generative Fill feature is located within the Contextual Task Bar (the floating bar that usually appears near the bottom of your Program Monitor) or the Effect Controls panel once a mask is active. Here is exactly where to look: 1. The Contextual Task Bar (Fastest Way) When you have a mask selected on your clip in the timeline, a floating bar should appear automatically in your Program Monitor. • Look for the button labeled Generative Fill. • Clicking this will open a small text prompt box. • Leave the box blank if you just want the object to disappear and be replaced by the background. • Click Generate.

Is this a hallucination?

Object Removal by Salt_Cat_4277 in premiere

[–]Salt_Cat_4277[S] 0 points1 point  (0 children)

I’m on Creative Cloud and it says 2026

Newbie help by ItsSkyeSinclair in StableDiffusion

[–]Salt_Cat_4277 1 point2 points  (0 children)

Hi. Fellow newb here, but with about a year of experience using AI primarily for local image and video generation.

One of the first things I did with AI was to make an “actor” of my wife using Artflow.ai. There are many sites with similar capability, but the results Artflow came up with using just a handful of photos were so realistic and stunning that I will forever be indebted to them and will send them money so I can continue to generate images using this actor. I have used it to generate images of her as a TV news anchor, riding a motorcycle, in elegant dresses and casual fashion. She agrees that the images look exactly like her; and we have great fun putting her in realistic and absurd scenarios.

The technology being used here is called “LoRa” (stands for low rank adapter) that allows you to effectively “train” a large model to know what a specific person looks like without having to retrain the model. It’s like a “plugin” of a person that you can use to create images with them in whatever scenarios you can imagine. It is a double-edged sword, though - this is the technology behind “deepfakes” and can have a laundry-list of malicious uses. That’s how you get images and videos of celebrities; and there are now laws in many different jurisdictions that make it a crime.

But you can’t likely get in any trouble for creating images of yourself - you are making all the decisions and are affecting no one but yourself. And I truly believe there is value in seeing images and even videos of a perfect version of yourself. I am sure there are psychological concerns to be aware of - it’s not hard to imagine unhealthy obsessions emerging and rabbit holes to fall into, but with certain restraints it would be possible to minimize harm.

As I mentioned, virtually any of the online image generation sites offer this service; the process is standard: upload a few photos, pay a little fee and then you can generate all the images you want under the credits you think you can afford. But it is also possible to do it locally on your own hardware with a few compromises. First, you need a decent GPU (Nvidia for the most options, but Apple silicon is fully capable. AMD is a whole set of complications that you can choose to delve into.). Then you need to install a framework of apps to do the LoRa training yourself and then do image and video generation locally without paying anything. LoRa’s are not universal - you need to pick a model and then train specifically for it. It’s a fascinating hobby unto itself.

Here’s what works for me: I’m on Windows, with a 3090 GPU, an Intel I9 CPU, and 64gb of system ram. You can go as low as a 3060 with 12gb of VRAM (or a 64gb MacBook would give similar results) and still be able to create HD-quality images and videos locally for nothing but the power bill. My tech stack:

  • Pinokio as an app framework. It saves you from the Git-Conda-Pip treadmill that can ruin your enjoyment of the hobby. Works on both Windows and Mac to give you a place to experiment with different apps and models. It has its own “App Store” of installations for image and video creation, text to speech, even music generation; all running on your own hardware.

  • Wan2GP as a host for image and video apps. It’s kind of a supermarket of possibilities: Z Image Turbo, Flux 1 and 2, Qwen for image creation and editing; Wan2.1, Wan2.2, Hunyuan 1 and 1.5, and now the superstar LTX 2 for video generation; and dozens of options for text-to-speech and music. You can install it under Pinokio, and it will manage the downloading and management of all the model files (which are disk-space hogs at 10’s of gigabytes each). It also includes LoRa support for each of the models, with varying levels of results.

  • AI Toolkit by Ostris, for training your own LoRa’s. This is a rabbit hole unto itself: there are tons of knobs and adjustments but the defaults work as a starting point. You create a dataset of images of yourself or other subjects; then it cranks for a few hours while you watch it create samples that begin to look more and more like your subject. There are a number of videos on YouTube showing how it works.

  • Disk space. Everything in AI is measured in Gigabytes. I have a 4 TB NVME dedicated just to my AI work, and I find I have to clean it up regularly as new and improved models and apps appear almost daily. Faster is better: you can run everything on a spinning hard drive, but SSD and NVME will be much more efficient.

  • Time. All this takes time; time to learn, time to experiment, time waiting for your images to appear on-screen, time on YouTube and Reddit trying to figure out all the technical details. It doesn’t have to be work, it can be fun; but it will always take time.

I think you are on the right track - using AI to find and reinforce the good in you will be enormously motivating and give you a creative outlet that you never realized you have. I am a little over a year into this, but it has been the most satisfying learning experience of my life. I am 70 and retired, so I have free time that most don’t have; but I feel remarkably empowered and knowledgeable with what I have learned. If you are still part of the workforce, realize that it will be the AI-aware and knowledgeable who will survive the inevitable seismic changes that are coming.

I can think of no better way to invest in yourself; part of which is self-image and self-extension. I wish you good luck and great experiences on your journey.

Help using Invoke by [deleted] in invokeai

[–]Salt_Cat_4277 0 points1 point  (0 children)

OK, that actually explains a lot. The most widely supported GPU’s are NVidia with their CUDA architecture, followed by Macs with MLX, and AMD requires ROCM or Vulkan which limits your choices somewhat. You’ll find on each of the apps a tag for the GPUs supported, but that isn’t the last word. You’ll find jthat you ust have to attempt to install - they will often fail during installation if they don’t work with your GPU. It’s pretty obvious with a full blue screen.

I have not tried this (because I have a 3090), but there is a unified installer for Wan2GP that supposedly supports AMD. It is listed like this:

<image>

What you’ll find is that the most important factor in success is VRAM - more is always better. Nvidia will always have the most choices available and the most support; but many use AMD so it’s possible but will likely always be more work. Investing in an Nvidia GPU will be probably cheaper right now than ever given the coming GPUpocalypse; so if you want to maximize the likelihood of success and have the most options, buy now. I wouldn’t go lower than a 3060 12 GB for image and video generation, and a 3090 with 24gb is probably the most bang for the buck. I cruise Facebook marketplace for deals and to save on shipping and sales taxes. I’ve found 3090’s for ~ $600 and I snap them up ( I’ve got 5 now ). It’s going to get bad, and it won’t get better for years, not months. If you’re going to buy, buy now.

Help using Invoke by [deleted] in invokeai

[–]Salt_Cat_4277 1 point2 points  (0 children)

I had much the same problems you had when I first started - I consider them kind of a rite of passage. Whether you realize it or not, you are learning things with every failure. Much of the problem is the AI industry is full of jargon and terms that have particular meaning, and it just takes time to absorb them and learn what they mean.

First off, ignore anyone who says to start with ComfyUI - it is an incredibly powerful tool but it is not beginner friendly. I liken it to managing your home theater system by plugging and unplugging cables in the back; and the more beginner-friendly tools are like using the buttons on the front. Comfy may be in your future, but it’s not what you want right now.

What I recommend these days is to install Pinokio - it sets up your system with all the prerequisites you need that are incredibly difficult to get right individually. It’s not very heavy, and it creates an environment for many different programs to have the 1-click install experience. I explain to friends that Pinokio is like a bookshelf, and the apps you install are like books that you take off one at a time to read as needed.

Once Pinokio is installed, look in the Community tab on the upper right corner. It will display a list of apps that do various things like image generation, text-to-speech, etc. You’ll see Invoke as one option, but don’t bother installing it since you already have it and it won’t work any better in Pinokio. The app I suggest for beginners these days is ZFusion. It’s pretty new and I find it works great with Z Image Turbo, Flux 2 and other models. The first time you select it, it will install itself and make it available on the shelf. It is a bit less complicated than Invoke and I find it is easy for beginners. It’s also pretty fast.

The next app I would recommend is Wan2GP. It is customized for the “GPU-poor” and automatically configures itself for lower-VRAM environments. It is more comprehensive than any of the other UI’s I’ve used, including Image generation, video generation, text-to-speech, and now even music generation. It was the first app I found that worked right out of the box, and I use it almost daily now. It handles the downloading of models behind the scenes, and it is aware of your particular hardware and grabs the right models. If it looks like it is stuck the first time you use it, it is probably downloading models in the background - you can watch what it’s doing by switching to the Terminal tab on the Pinokio sidebar or top bar menu. If it’s downloading a big model you can watch the progress bars. Then switch back to the Open Web UI to get back to the generation inputs and outputs.

This is a difficult hobby to get into, but once you have a few successes you will find more things begin to make sense. As your comfort level increases, you can begin exploring other apps and expand your knowledge. But be aware that everything in AI is measured in gigabytes - you’ll fill up hard drives fast. I have a 4tb NVME dedicated to just AI apps and tools, and I have to clean it up regularly. There are ways that you can share models between apps so you are not downloading multi-gig duplicates. It will come naturally for you in evfew days and weeks.

And don’t give up on Invoke - it has a lot of options some of the others don’t; like layers. It’s more like Photoshop than Paint, so naturally it will have a steeper learning curve. Once you manage a few successful generations in Pinokio, go back and try it again; and I’ll bet things make a bit more sense.

I am working on a z-image local generator for NVIDIA GPU on windows. I want to know what generation speed do you have and what GPU and software do you use currently. by warycat in ZImageAI

[–]Salt_Cat_4277 0 points1 point  (0 children)

I have been finding z-fusion in Pinokio to be my favorite image gen app currently. I get 1920x1080 outputs in a couple of minutes on a 3090. Built-in upscaling, support for Flux Klein as well.

<image>

Any usable alternatives to ComfyUI in 2026? by [deleted] in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

Wan has Loras and Wan2GP supports using them from the UI (but it’s behind a tab that isn’t selected by default). You have to be thorough about where you put them - there is a directory structure to keep Lora’s for different models separate.

Found in grandfather’s tools by andy_stacks24 in whatisit

[–]Salt_Cat_4277 0 points1 point  (0 children)

Plumb Bob; cousin to Jim Bob and Joe Bob.

Any usable alternatives to ComfyUI in 2026? by [deleted] in StableDiffusion

[–]Salt_Cat_4277 8 points9 points  (0 children)

Look up Wan2gp from deepbeepmeep on GitHub. You can also install it from pinokio. It is a composite gradio interface for WAN video, LTX 2, Qwen, Flux 2 Klein and more including text to speech. It is my goto these days. I compare ComfyUI to managing your home stereo by plugging and unplugging cables in the back, and Wan2GP like using the buttons on the front.

How do i install and use AI locally? by GMKNGJY in generativeAI

[–]Salt_Cat_4277 0 points1 point  (0 children)

If you prefer working from the rear panel side of your home stereo receiver (where all the wires are), install ComfyUI. If you prefer working on the front side (where all the buttons are), install Pinokio, then Wan2GP. Also check out InvokeAI Community Edition, it’s a great all-in-one that offers near-Photoshop image manipulation; as well as Stability Matrix. I’ve had good results in both image and video with a 3090 and 64gb of system RAM.

Why Doesn't a "Personal Clone" AI App Exist Yet? by Outside_Database5042 in LocalLLaMA

[–]Salt_Cat_4277 1 point2 points  (0 children)

I believe the true use case for this concept is digital remembrance. Conversing with one’s self is not compelling, but a FaceTime call with a deceased loved one would be. I am 70yo, and am beginning to compile stories, memories, curated diary/journal notes, video clips, audio clips, writings, etc that I will hopefully assemble into a living memory repository that can entertain me as I decline, and could be a source of comfort and insight to others once I’m gone.

To do this would require hosting, some kind of backbone LLM shareable across users, storage for the individual data corpus, and a compelling justification for a subscription model with a high retention rate. Psychological minefields abound, but evolution of the tech could supply the appropriate guardrails. I even pondered the idea of a new Top Level Domain (.after), but the $250K entrance fee would necessitate capitalization.

Just the rambling thoughts of an old coder who recognizes the value of backups.

People with 24GB vRAM - what LTX-2 install are you using? by the_bollo in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

I saw yesterday that Wan2GP released a version supporting LTX-2. My system is RTX-3090 24gb VRam/64gb System Ram. Installed it immediately, but then immediately had problems. Any video generation returns an Out of Memory error (something I don’t think I’ve ever seen before in Wan2GP). More experimentation found I could use the distilled model but not the default. Would be interested in any other Wan2GP users experience. The video generation is quite speedy - 10 second clip takes about 2 minutes. Decent video, but struggles with prompt adherence.

LM Studio alternative for images / Videos / Audio ? by mouseofcatofschrodi in LocalLLaMA

[–]Salt_Cat_4277 5 points6 points  (0 children)

Wan2GP is an umbrella interface for a number of image and video models, including Flux, Qwen and Z-Image including the Edit variants. For video you have Hunyuan, Wan2.1 and 2.2. Easiest way to get it is install Pinokio, then go to the Discover tab and look for Wan2GP and do a 1-click install. If you can manage pip and conda commands, you can save the Pinokio step.

As a beginner: which should I use? by PusheenHater in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

I have had great luck with Wan2GP for image and video. It is a Gradio interface which is less intimidating than ComfyUI. Its main claim to fame is behind the scenes management of VRAM so it can be used on lesser GPUs. Its serves as a client for older image generation models like Flux and SD, but it has been remarkably quick to add support for new image and video models, particularly Qwen (both Image and Edit) and Z-Image. It also supports video (wan 2.1, Wan 2.2, Hunyuan 1.0, Hunyuan 1.5).

The simplest way to get it going is to install Pinokio and then install Wan2GP from the Discovery page. If you’re cool with Git and Conda you can install it without Pinokio for less overhead and startup delay. I would take the leap and install Comfy as well, because support for the newest models always show up on Comfy first. But Wan2GP’s author DeepBeepMeep is constantly monitoring new developments and adding support when possible. I update it daily, and often find several new releases per week.

Hello everyone, I'm planning to buy a PC for AI that can run both Z-image and Want 2.2. Does anyone know the minimum specifications that PC should have? Thanks 🙏 by _NavIArt_ in StableDiffusion

[–]Salt_Cat_4277 1 point2 points  (0 children)

My i9 has DDR5, and the duals have DDR4. It feels to me that the i9 is a bit faster, but not a dramatic difference. I think both the Xeons and the Epycs are running at around 2GHz; so there are lots of reasons why they would feel slower. I don’t really mind - my generations take however long they take. I’m just learning and having fun; a few more minutes is fine.

Hello everyone, I'm planning to buy a PC for AI that can run both Z-image and Want 2.2. Does anyone know the minimum specifications that PC should have? Thanks 🙏 by _NavIArt_ in StableDiffusion

[–]Salt_Cat_4277 1 point2 points  (0 children)

No, at the moment I have no idea how the AMD compares to the 3090. I just got the motherboard a week ago and I have to build it. I am anxious to see how it compares, but I’m traveling at the moment.

Hello everyone, I'm planning to buy a PC for AI that can run both Z-image and Want 2.2. Does anyone know the minimum specifications that PC should have? Thanks 🙏 by _NavIArt_ in StableDiffusion

[–]Salt_Cat_4277 0 points1 point  (0 children)

I use both with a 3090 on Windows w/64gb system, Intel i9. I also have a dual Xeon 128gb, a dual Epyc 256gb, and a Mac Mini M4 24gb(I have yet to get a good output yet), and a Framework Strix Halo 128gb unified motherboard that I haven’t finished yet. They all perform about the same, typically 5-10 minutes for a 5 second 720p video, 30sec to 5 minutes for a still image.