Draw Things
AppFreeNative Apple app for local AI image generation with Metal acceleration.
Capabilities14 decomposed
local text-to-image generation with metal-accelerated inference
Medium confidenceGenerates images from natural language prompts by executing Stable Diffusion and FLUX models directly on Apple Silicon devices using Metal GPU acceleration, eliminating cloud dependency and network latency. Models are downloaded once and cached locally, enabling offline generation after initial setup. The Metal acceleration framework optimizes tensor operations and memory bandwidth for M-series chips, delivering generation times measured in minutes per image on consumer hardware.
Implements Metal GPU optimization specifically for Apple Silicon's unified memory architecture, avoiding generic CUDA/OpenCL abstractions and enabling efficient tensor operations on M-series chips without cloud offload. Local model caching and offline-first design eliminates network round-trips entirely, unlike cloud-dependent competitors.
Faster than cloud-based alternatives (Midjourney, DALL-E) by eliminating network latency and queue times; more private than cloud services by keeping prompts and generations local; cheaper than cloud APIs for high-volume generation, but slower per-image than optimized cloud inference.
lora training and inference on-device
Medium confidenceEnables users to train custom Low-Rank Adaptation (LoRA) modules locally on Apple Silicon devices by fine-tuning base models (Stable Diffusion, FLUX) on user-provided image datasets. Trained LoRAs are stored locally and can be applied during inference to customize model outputs without retraining the full base model. The training process uses gradient descent optimization on-device, with inference applying LoRA weights as low-rank matrix multiplications during the diffusion process.
Performs LoRA training entirely on-device without cloud upload, preserving data privacy and enabling immediate iteration. Uses Metal-optimized gradient computation for Apple Silicon, avoiding generic PyTorch/TensorFlow frameworks that would be slower on mobile devices.
More private than cloud LoRA training services (Replicate, Hugging Face) by keeping training data local; faster iteration than cloud services due to no upload/download overhead; less flexible than full fine-tuning frameworks (Kohya, ComfyUI) but more accessible to non-technical users.
multi-model support with seamless switching
Medium confidenceSupports multiple image generation models (Stable Diffusion, FLUX, and others) with UI-based model selection, enabling users to switch between models for different generation tasks without restarting the app. Each model is downloaded and cached separately, and the app manages model loading and memory allocation. Implementation uses abstraction layer for model inference to support multiple architectures.
Implements abstraction layer for multiple model architectures, enabling seamless switching without app restart. Local model caching allows users to maintain multiple models simultaneously without cloud dependency.
More flexible than single-model services (DALL-E, Midjourney) by supporting multiple architectures; more convenient than manual model switching in frameworks like ComfyUI; less specialized than model-specific tools but more versatile.
native ios/ipados/macos unified interface
Medium confidenceProvides native UI implementations across iOS, iPadOS, and macOS using platform-specific frameworks (SwiftUI, UIKit) rather than cross-platform abstractions, enabling optimized UX for each platform. The unified codebase shares inference logic while maintaining platform-specific UI patterns and capabilities. iOS/iPadOS versions leverage touch input and mobile-optimized layouts; macOS version uses keyboard shortcuts and desktop-optimized workflows.
Implements native UI for each platform (SwiftUI for macOS, UIKit/SwiftUI for iOS) rather than cross-platform framework, enabling optimized UX and performance. Unified inference backend shares code across platforms while maintaining platform-specific UI patterns.
More responsive and native-feeling than web apps or cross-platform frameworks (React Native, Flutter); better integrated with Apple ecosystem (iCloud, Photos app, etc.); less flexible than web-based alternatives for cross-platform access.
free tier with optional paid upgrades
Medium confidenceOffers free local image generation on Apple Silicon devices with limited cloud compute hours (Lab Hours), with optional paid tier (Draw Things+) providing higher cloud compute quotas and custom LoRA cloud inference. Free tier enables full local inference without payment; cloud features are optional and quota-based. Pricing model uses monthly Lab Hours allocation rather than per-request billing.
Implements freemium model with local-first approach, enabling full functionality without payment while offering optional cloud acceleration. Quota-based billing provides cost predictability compared to per-request cloud APIs.
More accessible than cloud-only services (Midjourney, DALL-E) by offering free local generation; more cost-predictable than per-request APIs by using monthly quotas; less transparent than subscription services regarding pricing and quota allocation.
app store distribution with direct download fallback
Medium confidenceDistributes the application through Apple App Store for iOS/iPadOS/macOS with direct download option as fallback when App Store is unavailable or inaccessible. App Store distribution enables automatic updates and seamless installation; direct download provides alternative installation path for users in regions with App Store restrictions or experiencing connectivity issues.
Provides both App Store and direct download distribution, offering flexibility for users in different regions or with different connectivity constraints. Direct download fallback ensures accessibility when App Store is unavailable.
More convenient than manual installation by offering App Store distribution; more accessible than App Store-only by providing direct download fallback; less flexible than open-source distribution but more secure with code signing.
controlnet-guided image generation
Medium confidenceApplies ControlNet conditioning to text-to-image generation, allowing users to guide model outputs using structural constraints (edge maps, pose skeletons, depth maps, etc.) provided as input images. ControlNet modules are loaded alongside base models and inject spatial conditioning into the diffusion process, enabling precise control over composition, pose, or layout without full inpainting. Implementation uses cross-attention mechanisms to blend ControlNet embeddings with text prompt embeddings during denoising steps.
Implements ControlNet inference on Apple Silicon with Metal optimization, avoiding cloud dependency for spatially-guided generation. Integrates ControlNet conditioning directly into the local diffusion pipeline rather than as a separate post-processing step.
More private than cloud ControlNet services by keeping reference images and outputs local; faster than cloud alternatives by eliminating network latency; less flexible than full ControlNet frameworks (ComfyUI, Automatic1111) but more accessible to non-technical users.
inpainting and selective region image editing
Medium confidenceEnables users to edit specific regions of images by masking areas and regenerating only masked regions using the diffusion model, preserving unmasked content. The infinite canvas feature allows expanding the image boundaries and filling new regions with model-generated content. Inpainting uses masked diffusion, where the model only denoises masked pixels while keeping unmasked pixels fixed, enabling seamless blending of edited and original content.
Performs masked diffusion inference locally on Apple Silicon, enabling fast iterative inpainting without cloud round-trips. Infinite canvas feature allows expanding image boundaries and filling new regions, not just editing existing content.
Faster than cloud inpainting services (Photoshop Generative Fill, Runway) by eliminating network latency; more private by keeping images local; less feature-rich than desktop editing software (Photoshop, GIMP) but more accessible and integrated with generation workflow.
style transfer and image-to-image transformation
Medium confidenceTransforms images into different artistic styles (animation, painting, sketch, etc.) by using the source image as a conditioning input to the diffusion model. The process encodes the source image into latent space and applies style guidance through text prompts or style presets, regenerating the image in the target style while preserving composition and content. Implementation uses image encoding followed by conditional diffusion with style-specific prompts or LoRA weights.
Performs style transfer locally on Apple Silicon using conditional diffusion with Metal optimization, avoiding cloud upload of source images. Integrates style presets and LoRA-based styles directly into the generation pipeline.
More private than cloud style transfer services by keeping source images local; faster than cloud alternatives by eliminating network latency; less flexible than full image-to-image frameworks (ComfyUI, Automatic1111) but more accessible to non-technical users.
image-to-video animation generation
Medium confidenceConverts static images into short video clips by generating frame sequences that extend or animate the source image content. The process likely uses frame interpolation or latent space animation to create smooth transitions between generated frames. Implementation details are undocumented, but the feature appears to create basic animations rather than full video generation from scratch.
Performs video generation locally on Apple Silicon without cloud dependency, though implementation approach is undocumented. Integrates video generation into the same interface as image generation, enabling seamless workflow from image to video.
More private than cloud video generation services by keeping source images and outputs local; faster than cloud alternatives by eliminating network latency; less capable than dedicated video generation models (Runway, Pika) but more integrated with image generation workflow.
model download and local caching management
Medium confidenceManages the download, storage, and lifecycle of image generation models (Stable Diffusion, FLUX, ControlNets, LoRAs) from Draw Things servers to local device storage. Models are downloaded once and cached locally, enabling offline inference after initial setup. The system tracks model versions, manages storage quotas, and provides UI for model selection and deletion. Implementation uses HTTP downloads with resume capability and local filesystem storage with version tracking.
Implements local model caching with offline-first design, enabling inference without cloud connectivity after initial download. Integrates model management directly into the app UI rather than requiring manual filesystem operations.
Simpler than manual model management in frameworks like ComfyUI or Automatic1111; more convenient than downloading models from Hugging Face manually; less flexible than custom model sources but more curated and optimized for Apple Silicon.
optional cloud compute offload with quota-based billing
Medium confidenceProvides optional cloud inference capability for users who want faster generation or higher quality outputs than local device allows, with quota-based billing using 'Lab Hours' currency. Free tier includes limited monthly Lab Hours; paid tiers (Draw Things+) increase quota. Cloud inference uses Draw Things' managed servers and supports custom LoRA inference, which is not available in free local tier. Implementation uses account-based authentication and cloud API for offloading generation requests.
Implements optional cloud offload with quota-based billing rather than per-request pricing, allowing users to control costs predictably. Integrates seamlessly with local inference, enabling users to switch between local and cloud generation in the same UI.
More flexible than cloud-only services (Midjourney, DALL-E) by supporting local generation; more cost-predictable than per-request cloud APIs by using monthly quotas; less transparent than cloud services regarding data handling and privacy.
try it on virtual fitting for apparel and characters
Medium confidenceEnables users to visualize apparel or character designs on virtual models or in different contexts by applying generated or uploaded designs to reference images. The mechanism is undocumented but likely uses image-to-image transformation or ControlNet-guided generation to adapt designs to different poses, body types, or character models. Allows rapid prototyping of apparel concepts without physical samples.
Integrates virtual fitting directly into the local image generation pipeline, enabling rapid prototyping without cloud dependency. Mechanism is undocumented but appears to use conditional generation to adapt designs to different models.
More private than cloud virtual fitting services by keeping designs local; faster than cloud alternatives by eliminating network latency; less specialized than dedicated fashion tech platforms (CLO, Browzwear) but more accessible and integrated with design workflow.
prompt engineering and generation parameter control
Medium confidenceProvides UI controls for fine-tuning text-to-image generation through prompt input and numerical parameters (guidance scale, sampling steps, seed, etc.). Guidance scale controls how strongly the model follows the text prompt; sampling steps control diffusion quality and speed; seed enables reproducible generation. Implementation uses standard diffusion parameter exposure through UI sliders and text input fields.
Exposes diffusion parameters directly in the UI with real-time feedback, enabling users to understand parameter effects without external documentation. Seed-based reproducibility enables iterative refinement of specific generated images.
More transparent than cloud services (Midjourney) regarding parameter effects; more accessible than command-line tools (ComfyUI, Automatic1111) but less flexible for advanced parameter experimentation.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with Draw Things, ranked by overlap. Discovered automatically through the match graph.
Llama 3.2 11B Vision
Meta's multimodal 11B model with text and vision.
FLUX.1-RealismLora
FLUX.1-RealismLora — AI demo on HuggingFace
Lensa
An all-in-one image editing app that includes the generation of personalized avatars using Stable...
ollama
Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.
FLUX-LoRA-DLC
FLUX-LoRA-DLC — AI demo on HuggingFace
flux-lora-the-explorer
flux-lora-the-explorer — AI demo on HuggingFace
Best For
- ✓macOS and iOS users with Apple Silicon devices (M1/M2/M3/M4 chips)
- ✓Privacy-conscious creators unwilling to send prompts to cloud services
- ✓Individual artists and designers prototyping concepts locally
- ✓Users in regions with limited cloud service availability
- ✓Individual creators building custom character or style models for personal use
- ✓Teams with proprietary training data unwilling to upload to cloud services
- ✓Artists iterating rapidly on custom model training with immediate feedback
- ✓Users in privacy-sensitive jurisdictions or organizations
Known Limitations
- ⚠Apple Silicon only — no Windows, Linux, or Intel Mac support documented
- ⚠Generation speed varies significantly by device model and available VRAM; no published benchmarks provided
- ⚠Models must be downloaded and stored locally before use; typical model sizes 2-7GB per model
- ⚠Maximum concurrent generations limited by device memory; no batch processing API documented
- ⚠Image resolution and quality constrained by available device memory; maximum resolution not publicly specified
- ⚠LoRA training time and resource requirements not publicly documented; likely hours to days depending on dataset size and device
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Native macOS and iOS application for running Stable Diffusion, FLUX, and other image generation models locally on Apple Silicon with Metal acceleration, offering LoRA support, ControlNet, inpainting, and optimized performance without cloud dependencies.
Categories
Alternatives to Draw Things
Are you the builder of Draw Things?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →