Agent SkillsAgent Skills
binhmuc

ai-artist

@binhmuc/ai-artist
binhmuc
22
3 forks
Updated 3/31/2026
View on GitHub

Write and optimize prompts for AI-generated outcomes across text and image models. Use when crafting prompts for LLMs (Claude, GPT, Gemini), image generators (Midjourney, DALL-E, Stable Diffusion, Imagen, Flux), or video generators (Veo, Runway). Covers prompt structure, style keywords, negative prompts, chain-of-thought, few-shot examples, iterative refinement, and domain-specific patterns for marketing, code, and creative writing.

Installation

$npx agent-skills-cli install @binhmuc/ai-artist
Claude Code
Cursor
Copilot
Codex
Antigravity

Details

Path.claude/skills/ai-artist/SKILL.md
Branchmain
Scoped Name@binhmuc/ai-artist

Usage

After installing, this skill will be available to your AI coding assistant.

Verify installation:

npx agent-skills-cli list

Skill Instructions


name: ai-artist description: Write and optimize prompts for AI-generated outcomes across text and image models. Use when crafting prompts for LLMs (Claude, GPT, Gemini), image generators (Midjourney, DALL-E, Stable Diffusion, Imagen, Flux), or video generators (Veo, Runway). Covers prompt structure, style keywords, negative prompts, chain-of-thought, few-shot examples, iterative refinement, and domain-specific patterns for marketing, code, and creative writing. version: 1.0.0 license: MIT

AI Artist - Prompt Engineering

Craft effective prompts for AI text and image generation models.

Core Principles

  1. Clarity - Be specific, avoid ambiguity
  2. Context - Set scene, role, constraints upfront
  3. Structure - Use consistent formatting (markdown, XML tags, delimiters)
  4. Iteration - Refine based on outputs, A/B test variations

Quick Patterns

LLM Prompts (Claude/GPT/Gemini)

[Role] You are a {expert type} specializing in {domain}.
[Context] {Background information and constraints}
[Task] {Specific action to perform}
[Format] {Output structure - JSON, markdown, list, etc.}
[Examples] {1-3 few-shot examples if needed}

Image Generation (Midjourney/DALL-E/Stable Diffusion)

[Subject] {main subject with details}
[Style] {artistic style, medium, artist reference}
[Composition] {framing, angle, lighting}
[Quality] {resolution modifiers, rendering quality}
[Negative] {what to avoid - only if supported}

Example: Portrait of a cyberpunk hacker, neon lighting, cinematic composition, detailed face, 8k, artstation quality --ar 16:9 --style raw

References

Load for detailed guidance:

TopicFileDescription
LLMreferences/llm-prompting.mdSystem prompts, few-shot, CoT, output formatting
Imagereferences/image-prompting.mdStyle keywords, model syntax, negative prompts
Nano Bananareferences/nano-banana.mdGemini image prompting, narrative style, multi-image input
Advancedreferences/advanced-techniques.mdMeta-prompting, chaining, A/B testing
Domain Indexreferences/domain-patterns.mdUniversal pattern, links to domain files
Marketingreferences/domain-marketing.mdHeadlines, product copy, emails, ads
Codereferences/domain-code.mdFunctions, review, refactoring, debugging
Writingreferences/domain-writing.mdStories, characters, dialogue, editing
Datareferences/domain-data.mdExtraction, analysis, comparison

Model-Specific Tips

ModelKey Syntax
Midjourney--ar, --style, --chaos, --weird, --v 6.1
DALL-E 3Natural language, no parameters, HD quality option
Stable DiffusionWeighted tokens (word:1.2), LoRA, negative prompt
FluxNatural prompts, style mixing, --guidance
Imagen/VeoDescriptive text, aspect ratio, style references

Anti-Patterns

  • Vague instructions ("make it better")
  • Conflicting constraints
  • Missing context for domain tasks
  • Over-prompting with redundant details
  • Ignoring model-specific strengths/limits

More by binhmuc

View all
mobile-development
22

Build modern mobile applications with React Native, Flutter, Swift/SwiftUI, and Kotlin/Jetpack Compose. Covers mobile-first design principles, performance optimization (battery, memory, network), offline-first architecture, platform-specific guidelines (iOS HIG, Material Design), testing strategies, security best practices, accessibility, app store deployment, and mobile development mindset. Use when building mobile apps, implementing mobile UX patterns, optimizing for mobile constraints, or making native vs cross-platform decisions.

sequential-thinking
22

Apply structured, reflective problem-solving for complex tasks requiring multi-step analysis, revision capability, and hypothesis verification. Use for complex problem decomposition, adaptive planning, analysis needing course correction, problems with unclear scope, multi-step solutions, and hypothesis-driven work.

ai-multimodal
22

Process and generate multimedia content using Google Gemini API for better vision capabilities. Capabilities include analyze audio files (transcription with timestamps, summarization, speech understanding, music/sound analysis up to 9.5 hours), understand images (better image analysis than Claude models, captioning, reasoning, object detection, design extraction, OCR, visual Q&A, segmentation, handle multiple images), process videos (scene detection, Q&A, temporal analysis, YouTube URLs, up to 6 hours), extract from documents (PDF tables, forms, charts, diagrams, multi-page), generate images (text-to-image with Imagen 4, editing, composition, refinement), generate videos (text-to-video with Veo 3, 8-second clips with native audio). Use when working with audio/video files, analyzing images or screenshots (instead of default vision capabilities of Claude, only fallback to Claude's vision capabilities if needed), processing PDF documents, extracting structured data from media, creating images/videos from text prompts, or implementing multimodal AI features. Supports Gemini 3/2.5, Imagen 4, and Veo 3 models with context windows up to 2M tokens.

code-review
22

Use when receiving code review feedback (especially if unclear or technically questionable), when completing tasks or major features requiring review before proceeding, or before making any completion/success claims. Covers three practices - receiving feedback with technical rigor over performative agreement, requesting reviews via code-reviewer subagent, and verification gates requiring evidence before any status claims. Essential for subagent-driven development, pull requests, and preventing false completion claims.