Javascript Blocker Detected

Please temporarily pause or whitelist us to create an account

If you opened our link from an app, look for 3 dots on the top right to open our page in your browser. We don't display ads nor sell/share your data, but our website requires Javascript to work properly. Thanks! -Team Graydient

PirateDiffusion Guide

Pirate Diffusion

Is there a Checkpoint, Embedding, LoRA, or other model that you want preloaded?  Request a model install

pirate diffusion logo

Overview

Pirate Diffusion by Graydient AI is the most powerful bot on Telegram. It is a multi-modal bot, meaning that it handles large language models like Kimi, DeepSeek and Llama 3.3, tens of thousands of image models like HiDream, Chroma, FLUX 2, Stable Diffusion 3.5, AuraFlow, Pony, and SDXL and also video models like Wan 2.2, Hunyuan and LightTricks (LTX 2.3).

 

Incredible Value

Unlike other generative AI services, there are no “tokens” or “credits” required. PirateDiffusion is designed for unlimited use, royalty free, and it’s bundled with Graydient’s webUI. You can make unlimited AI videos, unlimited upscaled images, and so on.

Why use a bot?

It’s extremely fast and lightweight on mobile, and you can use it solo or in groups of strangers or with friends. Our community is making chat macros for powerfuly ComfyUI workflows, so you get the full benefit of a desktop rendering result from a simple chat bot. It’s insane. 

 

What else can it do?

 

Images, Video, and LLM chats. It can do just about anything. 

You can make 4k images in a few keystrokes without a GUI. You can use every major Stable Diffusion via chat. Some features require a visual interface, and it pops up a web page for those (Unified WebUI integration)

Create by yourself with your private bot, or join groups and see what other people are making. You can also create themed bots like /animebot which connect our PollyGPT LLMs to StableDiffusion models, and chat with them to help you create! Create a workflow that’s totally unique to you using loadouts, recipe (macros) widgets (visual gui builder) and custom bots. It does a lot!

Origin Story

The name Pirate Diffusion comes from the leaked Stable Diffusion 1.5 model in October 2022, which was open sourced. We built the first Stable Diffusion bot on Telegram and thousands of people showed up, and that’s how we started. But to be perfectly clear: there’s no piracy going on here, we just loved the name. Still, enough people (and our bank) said the name was a little much, so we renamed our company to “Graydient” but we love Pirate Diffusion. It attracts fun, interesting people.

Quick start – Activate and Skills

Talk to @piratediffusion_bot in Telegram.  Type /start and then type /debug

If it also asks you for your email address, please enter it following the commands on the screen.  If the debug says “free-free” it means your bot is not activated.  Please contact our technical support team (we have live chat on Telegram or email) to activate the bot for you.  This will also lift the NSFW restrictions.

SKILLS COMMAND

Next, let’s create an image or video with a Skill command.  A skill is a universal command that makes prompting easy.  

Begin your request with two slashes, like this: 

//

For example:

The // command can also generate videos, edit images, and more.  If you’re in a hurry and don’t want to read the rest of the manual, you’re done!  But to make the most of out of our software, we encourage you to keep going.

Deep dive into skills and customing them

You can customize what the // command does by learning the Skills feature. here’s the full guide.

Manual video model selection: Text-to-Video 

Step-by-Step tutorial & examples: https://youtu.be/21kBNiR1Jro

First, browse our workflows page and find the text to video short names.  The popular ones are:
  • LTX 2.3 – the newest video model creates SFW video with cohesive speech and text
  • WAN 2.2 – the current top open source video model, write a simple prompt and it happens! We have hundreds of WAN Loras (mini models) for specific poses and effects. It’s quite good.
  • Boringvideo – creates lifelike ordinary videos that look like they came from an iPhone
  • HunYuan – three types, look at the bottom of the workflow page to find them. Hunyuan creates the most realistic videos. The higher the “Q” number the greater the quality is, but the shorter the videos are.
  • Video – the ones simply called Video use older versions of LTX, great for 3D cartoons
 
/wf /run:video-wan22 cinematic low angle video of a ronald mcdonald clown eating a square hamburger, the restaurant ((sign text says "Wendys")), ronald mcdonald's clown costume has red afro hair and a red nose with white face paint, the restaurant is brown, the burger is pointy and square, the background has blur bokeh and people are walking around /length:120
 

After the video is created, you can prompt the next section and keep expanding the video.  

Note: The Wan model is designed to loop at around 161 frames, so a safe option is around 120 frames, unless you want to see the animation loop and repeat itself.

/wf /run:extend-wan22 he turns around and runs into the building /length:120

You can do this multiple times and keep extending the video (until it doesn’t fit on video memory). 

/wf /run:mmaudio funny music and clown voices, burger eating sounds, cartoon antics

You can add sounds and voices to the video as well with the MMaudio workflow.  It’s admittedly hard to control, but is good at sound effects.

/wf /run:video-upscale

Then upscale the final result to smooth frames and increase the resolution

PirateDiffusion supports WAN, Hunyuan and LightTricks / LTX videos and loras!  We are adding multiple video models and you can use them unlimited as part of our service, along with unlimited image and unlimited lora training.

In LTX, the structure of the prompt matters a lot. A short prompt will result in a static image. A prompt with too many actions and instructions will cause the video to pan to different random rooms or characters.

Best Practices: How to make your images move cohesively
 
We recommend a prompt pattern like this:
  1. First describe what the camera is doing or who its following. For example, a low angle camera zoom, an overhead camera, a slow pan, zooming out or away, etc.
  2. Next describe the subject and one action they are performing onto what or whom. This part takes practice!  In the example above, notice that Ronald eating the burger came after the camera and scene setup
  3. Describe the scene.  This helps the AI “segment” the things that you want to see. So in our example, we describe the clown costume and background.
  4. Give supporting reference material.  For example, say “This looks like a scene from a movie or TV show”
  5. You can specify a lora to control the art direction or character likeness. Add this to the end of the prompt like <move-enhancer-huny>

Image-to-Video

You can upload a photo and turn it into a video. There isn’t just one command – look at the “animate” workflow series to use different kinds of AI models. Try different models and prompt strategies to find the one that works best for your project, or look in the PLAYROOM channel in PirateDiffusion to see what others have created with them.

The aspect ratio of the video will be determined by the image that you upload, so please crop it accordingly.

To do this, first paste the photo into the chat and click “Reply” as if you’re going to talk to the photo, and then give it a command like this:

/wf /run:animate-wan22 a woman makes silly faces towards the camera
or try one of the other workflows like LTX 2.3 and HunYuan and SkyReels:
/wf /run:animate-ltx23 camera video that slightly changes the angle, focused on a lovely girl smiling and looking at the camera, she looks curious and confident while maintaining eyes on the viewer, her hair is parted, she sits in front of a bookshelf and peeping gremlin eyes behind her, she is relaxing vibe

Wan 2.2 has the largest library of LoRA poses and characters, and seems to be best for naughty bits.

LTX 2.3 has superior audio, but is more SFW. So choose the right tool for the job.

More video workflows

We have many video (text-to-video) and animation (image-to-video) as well as video-to-video workflows.  Login to your web browser with the /webui command and click workflows to explore them.  If you don’t know your webui password, set one with /community /password:whatever in Telegram, then refresh the web browser to enter it in.

 

Special parameters:

/length = length of the video in frames.  Safe settings are 89, 97, 105, 113, 121, 137, 153, 185, 201, 225, 241, 257.   More is possible but unstable, depending on the AI model.  They are all different sizes, some use more VRAM.

/fps = frames per second.  24 is recommended.  Turbo workflows run at 18 fps but can be changed.  above 24 is cinematic, 30fps looks more realistic.  60fps is possible at low frames but it looks like chipmunk speed.

An amateur mistake is to go straight into the highest settings. Instead, work at 12 FPS and when you like a video, reply to it with the /workflow /run:video-upscale to get higher resolution and framerate.  If you don’t do it this way, your videos will be shorter and you won’t be able to upscale.  Remember that you can extend videos over and over, add audio later, then upscale them.  Do this in that order.

 
Try a new recipe! Type /render #quick and your prompt
Tip: Use /render #quick - a macro to achive this quality without typing negatives

Guidance (CFG)

The Classifier-Free Guidance scale is a parameter that controls how closely the AI follows the prompt; higher values mean more adherence to the prompt. 

When this value is set higher, the image can appear sharper but the AI will have less “creativity” to fill in the spaces, so pixelation and glitches may occur. 

A safe default is 7 for the most common base models. However, there are special high efficiency models that use a different guidance scale, which are explained below.

SYNTAX

/render <sdxl> [[<fastnegative-xl:-2>]]
/guidance:7
/size:1024x1024
Takoyaki on a plate

How high or how low the guidance should be set depends on the sampler that you are using. Samplers are explained below. The amount of steps allowed to “solve” an image can also play an important role.

 

Exceptions to the rule 

Typical models follow this guidance and step pattern, but newer high efficiency models require far less guidance to function in the same way, between 1.5 – 2.5.  This is explained below:

High Efficiency Models

Low Steps, Low Guidance

Most concepts require a guidance of 7 and 35+ steps to generate a great image. This is changing as higher efficiency models have arrived.

These models can create images in 1/4 of the time, only requiring 4-12 steps with lower guidance. You can find them tagged as Turbo, Hyper, LCM, and Lightning in the concepts system, and they’re compatible with classic models. You can use them alongside Loras and Inversions of the same model family. The SDXL family has the biggest selection (use the pulldown menu, far right). Juggernaut 9 Lightining is a popular choice.

Some of our other favorite Lightning models are <boltning-xl> and <realvis4light-xl> which look great with a guidance of 2, steps between 4-12, and Refiner (no fix) turned off. Polish it off with a good negative like [[<fastnegative-xl:-2>]].  Follow it up with an upscale, and the effects are stunning!

Look into the notes of these special model types for more details on how to use them, like Aetherverse-XL (pictured below), with a guidance of 2.5 and 8 steps as pictured below.

VASS (SDXL only)

Vass is an HDR mode for SDXL, which may also improve composition and reduce color saturation. Some prefer it, others may not. If the image looks too colorful, try it without Refiner (NoFix)

The name comes from Timothy Alexis Vass, an independent researcher that has been exploring the SDXL latent space and has made some interesting observations. His aim is color correction, and improving the content of images. We have adapted his published code to run in PirateDiffusion.

/render a cool cat <sdxl> /vass

Why and when to use it: Try it on SDXL images that are too yellow, off-center, or the color range feels limited. You should see better vibrance and cleaned up backgrounds.

Limitations: This only works in SDXL. 

 
 

More tool (reply command)

The More tool creates variations of the same image

To see the same subject in slightly different variations, use the more tool. 

DESCRIBE A PHOTO

Updated!  There are now two modes of describe: CLIP and FLORENCE2

Generate a prompt from any image with computer vision with Describe! It is a “reply” command, so right click on the image as if you were going to talk to it, and write

/describe /florence

The additional Florence parameter gives you a much more detailed prompt. It uses the new Florence2 computer vision model.  /describe by itself uses the CLIP model

Example

Launch widgets within PirateDiffusion