• Home
  • YOUR COURSE PAGES
  • เพิ่มเติม
    • Home
    • YOUR COURSE PAGES
  • Home
  • YOUR COURSE PAGES
AI Fundamentals

PROMPTING

What is it?

a prompt refers to the specific input or instruction given to an AI model to generate a desired response. A prompt can be a question, a statement, a few keywords, or a detailed instruction. The quality, clarity, and specificity of the prompt directly influence the quality of the AI's response.

NOT ALL PROMPTING IS THE SAME

For instance:

  • In text-based AI models like ChatGPT, a prompt could be a question like, "What is the capital of France?" or a more complex instruction such as, "Write a short story about a detective solving a mystery in a futuristic city."
  • In AI art generators like Midjourney, a prompt could be, "Generate an image of a sunset over a snowy mountain range with a lone figure standing on a cliff."


The better and clearer the prompt, the more relevant and accurate the AI’s output will be. In essence, a prompt serves as the guiding instruction that steers the AI towards generating the desired output.

Learn More

In the course we will use both text-based AI models for some tasks (like scripts, research and more); and we'll use AI image and video generators like Runway, Veo and Midjourney for images and video.

Check out the official prompting instruction pages from specific tools by clicking on the icon buttons below & scroll down for a condensed overview of each tool. 

Midjourney
Runway ML
Chat GPT

Prompting Guides by Tool Type

Below, you'll find tailored prompting guides organized by AI tool categories:

  • Scriptwriting & Idea Generation Tools
  • AI Image Generation Tools
  • AI Video Generation Tools
  • AI Audio Generation Tools

Each section includes specific tools and best practices for crafting effective prompts tailored to that tool's unique capabilities. These guides are designed to help you get the most out of each AI platform—whether you're generating a screenplay, creating visuals, producing video content, or composing audio.

Important Note: AI prompts will always generate a result—but not always the same one. Even identical prompts can yield different outcomes. That’s part of the creative process.

We encourage you to use these guides as starting points. Think of them as templates rather than strict formulas. Not every project requires every element of a detailed prompt. The best way to learn is through experimentation—try different approaches, refine your inputs, and see what works best for your unique goals.

Explore. Experiment. Create.

PROMPTS: Scriptwriting and Idea Generation AI Tools

click the drop down arrows to revel more

ChatGPT excels at generating creative text, dialogue, storylines, and detailed narrative scripts for video, audio, or written formats. It works especially well when you provide clear goals and context.


How to Use:

When prompting ChatGPT:

  • Provide clear context (e.g., topic, format, length).
     
  • Specify elements like tone, style, and audience.
     
  • Use specific instructions rather than vague requests.
     

You can also collaborate interactively — ChatGPT allows back-and-forth refinement, helping you revise and expand ideas quickly.



What Makes ChatGPT Different:

  • Conversational & iterative: You can revise ideas collaboratively in real time.
     
  • Versatile & creative: Handles everything from corporate training scripts to fantasy stories.
     
  • Style-aware: Capable of mimicking specific genres, formats, or voices if given proper guidance.


 

Prompting Tips:

Include:

  • Clear theme or concept
     
  • Target audience
     
  • Desired tone (e.g., serious, funny, cinematic)
     
  • Character roles (if narrative-based)
     

Avoid:

  • Overloading with jargon (unless the audience expects it)
     
  • Leaving out structure if you want something formatted (e.g., bullets, timestamps, scenes)



How to Prompt ChatGPT for Creative Video Work (2025 Edition)


What ChatGPT Excels At:

ChatGPT is highly effective at:

  • Generating creative text, dialogue, and storylines
     
  • Writing narrative scripts in a wide range of genres
     
  • Structuring video content (e.g., documentaries, shorts, explainers)
     
  • Brainstorming unique video concepts tailored to your audience


 

How to Prompt ChatGPT (Best Practices 2025)

What to Include in Your Prompt:

  • Format (e.g., script, outline, concept list)
     
  • Tone/Style (e.g., dramatic, lighthearted, educational)
     
  • Length or timing (e.g., “3-minute script,” “5-part series”)
     
  • Audience type (e.g., beginners, kids, tech-savvy adults)
     
  • Medium (e.g., YouTube, TikTok, podcast, short film)
     

What’s New in 2025?


  • Multimodal Prompting: You can now include images, voice recordings, and video frames as part of your prompts to help ChatGPT create context-aware outputs (especially powerful when used with tools like Frames, Veo, or Canvas).
     
  • Live Refinement: ChatGPT (with GPT-4o) can update scripts or structures in real time based on partial inputs (e.g., “Keep the intro but change the ending to a cliffhanger”).
     
  • Fewer Instructions Needed: Thanks to better memory and fine-tuning, you can use shorter prompts, and ChatGPT will still pick up on structure, pacing, and audience tone more reliably.


 

Prompt Examples (Updated)

Narrative Script Prompt Example:

“Write a 2-minute video script in the style of a motivational short film. A single mom working two jobs finally gets her nursing license. Use emotional dialogue, a cinematic arc, and a powerful voiceover.”
 

Mini-Series Structure Prompt Example:

“Outline a 4-part YouTube series called ‘Modern Cowboys’, covering how traditional ranchers are using AI, drones, and solar energy in 2025. Each part should include a topic title, scene structure, and key interview ideas.”
 

YouTube Idea Prompt Example:

“Suggest 5 video ideas for a Gen Z-focused channel that mixes AI tools with creative hobbies (e.g., music, writing, art). The ideas should be fun, engaging, and easy to film.”
 

Tool-Specific Prompt (e.g., for Veo + Frames):

“Create a prompt for Google Veo to generate a cinematic scene of a dusty Southern town at golden hour. Include a teenage girl in cowboy boots walking along train tracks, with voiceover narration about chasing freedom. Style: modern western, soft lens flares, melancholy tone.”
 

Voiceover & Visual Blend Prompt:

“Write a 1-minute voiceover script for a dramatic AI-generated short about loneliness in the digital age. The visuals should include a man in a dark apartment, neon lights from the city outside, and slow zooms. Add a poetic narration with a hopeful ending.”
 

Prompting Pitfalls to Avoid in 2025:


Overloading your prompt with too many requests in one sentence
Instead, break it into parts or use bullet points:

“Give me:
 A logline,

Scene outline,

Sample dialogue.”

 

Assuming ChatGPT will guess your format


Be explicit:

“Write it as a 3-act script with headings: ACT I, ACT II, ACT III.”
 

Final Tips

  • Use “Continue from here” or “Now revise the ending” for fast iteration
     
  • Tell ChatGPT where the content will live (TikTok, Reels, short film — it matters!)
     
  • Ask it to reframe ideas for different platforms if you want to repurpose content
     
  • If using tools like Veo or Runway, you can even prompt ChatGPT to write prompts for those tools
     

Try It Yourself:

 https://chat.openai.com
Or use the ChatGPT iOS/Android app for on-the-go idea generation


Gemini specializes in generating multi-modal content, combining visuals with textual explanations.


Differences:

  • Use Gemini when you need visual structure, presentation-ready content, or are working in Google tools.
     
  • Use ChatGPT when you want story-rich dialogue, flexible script shaping, or deep narrative control.
     

Many creators benefit from using both in tandem — Gemini to plan visual layout and structure, ChatGPT to refine the tone and storytelling.



What Gemini Excels At (as of 2025):


Gemini is Google’s multimodal AI, designed to seamlessly generate content that blends visuals, text, voice cues, layout ideas, and media structure. It’s especially powerful when working on:

  • Educational videos
     
  • Product explainers
     
  • Visual-first content like YouTube, Reels, and Slides
     
  • Google ecosystem integrations (Docs, Slides, YouTube, Drive)


 

How to Prompt Gemini (Best Practices in 2025)

Include in Your Prompts:

  • Text + Visual Intent (e.g., “show melting ice” or “cut to drone shot”)
     
  • Aesthetic preferences (e.g., cinematic, minimalist, retro)
     
  • On-screen elements (text overlays, transitions, graphs, cutaways)
     
  • Audio/Voiceover tone (serious, warm, Gen Z-friendly, etc.)
     
  • Target platform (YouTube, Shorts, Slides, etc.)
     

What’s New in Gemini (2025 Updates):

  • Integrated Multimodal Context: Gemini now allows image + text + layout input, meaning you can upload a slide or sketch and ask Gemini to build a video or storyboard from it.
     
  •  Prompt Memory Across Tools: When used inside Google Workspace, Gemini retains context across Docs, Slides, and Sheets — so you can build multi-format content more fluidly.
     
  • Smarter Visual Guidance: Gemini is now more accurate at responding to prompts like “show data as a line graph over a cityscape background” — helpful for explainer or documentary-style videos.
     
  • Storyboard Mode (beta in some tools): Gemini can now generate frame-by-frame visuals with script guidance, great for pre-visualizing video sequences.
     


Prompt Examples (Updated for 2025)


Script + Visual Cues Prompt Example:

“Write a 2-minute educational video script explaining the causes of wildfires. Include voiceover narration, on-screen bullet points for key facts, visual cues for drone footage of forests, animated fire maps, and calming background music.”
 

Video Structure Prompt Example (Business):


“Outline a 4-part product demo video for a smart home device. Each part should include:
 Intro problem (30 sec),

Feature demo (1 min),

Testimonials (45 sec),

Call to action (15 sec).
Include visual transitions, overlay text, and suggested B-roll footage.”

 

Idea Generation Prompt Example (Fitness Channel):

“Generate 5 visual-first video ideas for a women’s beginner fitness channel. Each idea should include a title, visual hook, camera angle suggestion (e.g., top-down, side profile), and the type of background music or tone to use.”
 

Visual-Heavy Explainer Prompt Example:

“Create a 60-second Instagram Reel script about plastic waste in oceans. Use punchy stats in large on-screen text, B-roll of marine life and ocean pollution, and end with a bold CTA. Keep the tone urgent but hopeful.”
 

Data + Visual Presentation Prompt Example (Slides or Shorts):

“Design the script and visual plan for a short video based on this Google Sheet data (insert link). Include a title card, transitions, animated bar chart overlays, and a summary screen with key takeaways. Intended for LinkedIn.”
 

Character-Driven Scene Prompt Example (Narrative Video):

“Write a dramatic 1-minute script for a short film where a country girl confronts her brother about leaving the family farm. Set the scene at dusk with warm lighting. Include facial reaction cues and soft ambient music.”
 

Visual Details Gemini Responds Well To:

  • Camera angles: “top-down,” “close-up,” “wide drone shot”
     
  •  Aesthetics: “vintage country,” “clean startup style,” “gritty realism”
     
  • On-screen text ideas: “show stat: 70% of waste is unrecycled”
     
  • Transitions: “fade into sunrise,” “hard cut to crowd cheering”
     
  • Layout/scene planning: “Split-screen of two characters walking different paths”
     

Avoid:

  •  Ambiguous phrases like “cool visuals” or “nice transition” — be specific (e.g., “pan up over cornfield at golden hour”)
     
  •  Telling Gemini to “generate everything” at once without breaking down structure
     
  •  Vague audience cues like “for people” — instead say “for Gen Z music lovers” or “for parents of toddlers”
     

⚙Integration Tip:


When using Gemini inside Docs, Slides, or YouTube Shorts Editor, you can:

  • Highlight your rough outline → Prompt Gemini to expand into a full script
     
  • Paste data → Ask Gemini to visualize it with suggested chart styles
     
  • Insert screenshots → Ask Gemini to rewrite overlay text or generate alt visuals
     


Final Thoughts:

Gemini thrives when you treat it like a creative director and editor in one — give it structure, visuals, and tone, and it can build out highly integrated video content.
 

Try Gemini:

 https://gemini.google.com/


Claude (by Anthropic) is known for its ability to create thoughtful, emotionally intelligent, and ethically grounded content. It thrives in projects that require:

  • Narrative depth
     
  • Character-driven storytelling
     
  • Exploration of social or moral themes
     
  • Structured, philosophical, or emotionally complex prompts
     

Claude is particularly effective for:

  • Drama shorts
     
  • Ethical debates
     
  • Human-centered documentaries
     
  • Empathy-first marketing scripts
     
  • Scripts for social issues or mental health topics


Key Differences: Claude vs. ChatGPT vs. Gemini (Creative Use)


Claude stands out for its emotionally rich, ethically grounded storytelling — perfect for drama, documentaries, or social issue content where nuance and reflection matter. In contrast, ChatGPT is more dynamic and collaborative, excelling at fast-paced scriptwriting, genre work, and dialogue-heavy content. Gemini is the most visual of the three, ideal for creators working on educational, product-based, or multimedia content that requires clear visual structure, transitions, and integration with Google tools.


Claude (Anthropic):

  • Specializes in emotionally intelligent, ethically-aware storytelling
     
  • Ideal for drama, reflective narratives, and social issue content
     
  • Excels at long-form continuity and character-driven arcs
     
  • Less focused on visuals; stronger on themes, dialogue, and nuance


 

2025 Updates: What's New in Claude

  •  Better Context Awareness: Claude now handles very long prompts (up to 200K+ tokens), meaning it can retain deep narrative context and maintain continuity across scenes or episodes.
     
  •  Improved Character Psychology Modeling: Claude understands emotional arcs, motivations, and dilemmas in more human-like ways, making it ideal for screenwriting and reflective storytelling.
     
  •  More Global + Culturally Aware: The 2025 version of Claude is better at reflecting diverse voices, backgrounds, and ethical viewpoints, especially useful in social justice or documentary contexts.


 

How to Prompt Claude (Best Practices)

Include:

  • Central themes (e.g., grief, digital identity, addiction)
     
  • Emotional turning points or internal conflict
     
  • Social or cultural framing (who it’s for, what matters)
     
  • Character roles + brief backgrounds
     
  • Tone/style (e.g., reflective, raw, poetic, grounded)
     

Avoid:

  • Overly vague prompts (“write a video about friendship”)
     
  • Asking for purely visual concepts without context (Claude doesn’t prioritize image generation)
     
  • Requests with no emotional or social grounding


 

Prompt Examples (Updated for 2025)

Script Prompt Example:

“Write a 5-minute short film script about a refugee father trying to reconnect with his daughter over video calls while working abroad. Show emotional distance, cultural tension, and a hopeful resolution.”
 

Video Structure Prompt Example (Social Issue Doc):

“Outline a 3-part video documentary on the emotional toll of student debt in the U.S.
Structure:
 Introduction (facts & stats)

Personal stories (3 individuals)

Expert insight & conclusion.
Include tone, visual approach, and narration style.”

 

Video Ideas Prompt Example:

“Suggest four compelling video ideas exploring ethical challenges in biotechnology, aimed at Gen Z science enthusiasts. Include potential titles, central questions, and a real-world case to explore in each.”
 

Debate/Discussion Prompt Example:

“Create a script for a YouTube explainer that introduces the debate around AI-generated art and its impact on human creativity. Include perspectives from both sides, a neutral narrator, and a reflection at the end.”

When to Choose Claude

Use Claude when your creative work involves:

  • Real-world ethical dilemmas or social commentary
     
  • Emotionally complex storytelling
     
  • Philosophical themes, inner conflict, or psychological realism
     
  • Long-form narratives that require consistent tone and logic
     

Claude is less flashy, more literary. Think of it like your sensitive screenwriter friend who asks the deeper questions.


Try Claude:

 https://www.anthropic.com


What Copilot Excels At (2025)

GitHub Copilot, powered by OpenAI, is specifically designed for structured, logic-driven content, especially in coding, technical tutorials, developer documentation, and how-to walkthroughs.


This may not be the most chosen tool for most of you developing AI driven video content, but it maybe - depending on your topic. 


In 2025, Copilot continues to excel in:

  • Code-centric video scripting
     
  • Step-by-step technical tutorials
     
  • Clear and consistent formatting
     
  • Helping creators build instructional or productivity-driven content with minimal fluff
     

2025 Updates: What’s New in Copilot

  • Smarter Inline Suggestions: Copilot now recognizes video pacing and narration flow inside script-like documents (e.g., voiceover vs. code breakdown).
     
  • Tutorial Logic Awareness: It can now anticipate learning gaps, suggesting pauses, recap points, or alternative explanations — ideal for educational video creators.
     
  • IDE Integration for Creators: In tools like VS Code, Copilot can generate documentation, code comments, and script narration in tandem — speeding up video production for dev content.
     

How to Prompt Copilot (Best Practices)

Include:

  • Clear technical topic or project (e.g., “build a to-do app in JavaScript”)
     
  • Step-by-step format or video length
     
  • Voiceover/narration cues (optional but helpful)
     
  • Expected output (e.g., “Include code + brief explanations + summary”)
     
  • Platform style (e.g., YouTube, course module, LinkedIn Learning)
     

Avoid:

  • Abstract or open-ended narrative requests
     
  • Emotion-heavy drama or character storytelling
     
  • Purely visual storytelling without step-by-step instruction


 

Prompt Examples (Updated for 2025)

Script Prompt Example:

“Write a detailed 10-minute YouTube script for a beginner tutorial on building a password generator in Python. Include step-by-step code walkthroughs, beginner-friendly explanations, and occasional voiceover suggestions.”
 

Video Structure Prompt Example (Photography Channel):

“Outline a beginner video tutorial for photography basics. Divide it into 5 parts: camera types, framing techniques, lighting fundamentals, common beginner mistakes, and pro tips. Add short intros and transitions between segments.”
 

Video Ideas Prompt Example (Tech/Programming Channel):

“Give me 3 engaging YouTube video ideas for teaching Python basics to high school students. Include a working title, project overview, and key learning outcomes.”
 

Educational Video Module Prompt:

“Create an outline for a 3-part mini-course video series teaching HTML and CSS. Each video should build on the last, include coding examples, and end with a small challenge project.”
 

Explain Like I'm 12 Prompt:

“Write a script that explains how a ‘for loop’ works in Python to a middle schooler. Include visual metaphors, real-life comparisons, and simple language for voiceover.”
 

Copilot vs Claude vs ChatGPT vs Gemini (Creative Video Prompting)


Copilot is purpose-built for generating structured, step-by-step content, making it ideal for technical tutorials, code explanations, and instructional videos. It shines when you need clear logic, consistent formatting, and beginner-friendly walkthroughs. In contrast, ChatGPT is the most conversational and flexible, excelling in storytelling, dialogue, and iterative creative development — perfect for scripting narratives, YouTube videos, and character-driven content. Claude specializes in deep, ethically aware storytelling, and is best suited for projects involving emotional arcs, social issues, or philosophical themes. Finally, Gemini stands out for its tight integration of visuals and text, making it the go-to choice for creating visually structured content like explainer videos, product demos, or content destined for Google Slides, Docs, or YouTube Shorts.


Use Copilot when your video content is:

  • Educational or instructional
     
  • Centered around code, dev tools, or software workflows
     
  • For platforms like YouTube, Udemy, LinkedIn Learning, or company tutorials
     
  • Meant to teach, explain, or onboard
     

Avoid Copilot for emotional storytelling, drama, or conceptual ideation. It’s a technical co-pilot, not a screenwriter.


Try Copilot:

 https://github.com/features/copilot


Perplexity is an AI platform known for its fact-first, research-focused capabilities. In 2025, it's a go-to tool for creators who need:

  • Data-driven scripting
     
  • Accurate, up-to-date research
     
  • Credible citations and sources
     
  • Well-structured informational content
     

Think of Perplexity as your AI research assistant that can also draft compelling analytical narratives. It’s ideal for:

  • Documentaries
     
  • Educational explainers
     
  • Current events analysis
     
  • Economic, environmental, or scientific videos
     

2025 Updates: What’s New in Perplexity

  • Live Web Access: Perplexity now offers real-time, citation-backed search integration, meaning you get sources you can verify — ideal for factual content and journalistic work.
     
  • Auto-Referencing: When you prompt Perplexity for data or stats, it can automatically embed source links or citations, making it easier to build trusted content.
     
  • Improved Summary Logic: It now better condenses complex research into digestible sections, which is great for video segments and voiceover scripts.
     
  • Current Awareness Mode: Perplexity can tailor its response to recent news, global developments, or emerging topics — giving you timely content faster than most AI tools.
     

🛠How to Prompt Perplexity (Best Practices)

Include:

  • A specific topic or question (e.g., “What caused the 2008 housing crash?”)
     
  • Clear request for references or data
     
  • Intended format (e.g., “script for 5-min video,” “timeline-based breakdown”)
     
  • Desired depth or audience level (e.g., “explain for high school students”)
     

Avoid:

  • Highly speculative, fictional, or emotional storytelling
     
  • Ambiguous creative tasks without a research angle
     
  • Complex prompts without clarity on what facts or citations are needed
     

Prompt Examples (Updated for 2025)

Script Prompt Example:

“Generate a 5-minute video script explaining the key milestones in the development of quantum computing. Include three historical events, how they impacted the field, and references to key researchers.”
 

Video Structure Prompt Example (EV Industry):

“Outline a 6-part educational YouTube video on the global rise of electric vehicles.
Include:
 History of EVs

Modern tech breakthroughs

Market growth stats

Environmental impact

Government policies

Future projections.
Reference real-world data and reports.”

 

Video Ideas Prompt Example (Economics/Global Trends):

“Suggest three video ideas for a data journalism YouTube channel focused on global economic shifts. Each idea should include a working title, key stats to include, and suggested sources or regions to focus on.”
 

Explainer Prompt Example (Science/Tech):

“Write a short explainer video about CRISPR gene editing. Make it clear, factual, and suitable for college students. Include on-screen text suggestions for scientific terms, and mention 2 recent studies.”

Differences Overview

Perplexity is the best tool when accuracy, research, and real-time data are essential. It’s ideal for fact-based content like explainers, trend analysis, and educational videos. In contrast, ChatGPT is more narrative-driven, great for scripting stories or engaging dialogue. Gemini shines when visuals are core to your project — from slide-based videos to product explainers — while Claude offers emotional and ethical depth for reflective or issue-based storytelling. Copilot rounds out the group by excelling in instructional, code-focused content where step-by-step clarity is key.


Try Perplexity:

https://www.perplexity.ai


What Squibler Excels At 

Squibler is built for professional scriptwriters and long-form storytellers, offering specialized tools for outlining, formatting, and developing screenplays, episodes, and structured narrative arcs.

In 2025, Squibler is perfect for:


  • Film, TV, and YouTube scriptwriting
     
  • Structuring multi-part series
     
  • Designing character arcs, worldbuilding, and act breaks
     
  • Refining existing story drafts or idea templates
     

It’s like a digital screenwriting room with AI support baked in — focused more on structure and formatting than raw idea generation.


2025 Updates: What’s New in Squibler

  • Advanced Screenplay Formatting: Squibler’s AI now understands scene headers, dialogue blocks, stage directions, and pacing even better, aligning with industry standards.
     
  • Multi-Scene Planning Tools: You can now generate or refine scene-by-scene breakdowns with beats, transitions, and character focus areas — great for writers’ rooms or episodic planning.
     
  • Story Engine Expansion: The “Story Engine” now includes templates for genre-specific outlines (e.g., thriller, romance, sci-fi) that can be customized via AI prompting.
     
  • Dynamic Story Templates: 2025’s update allows you to prompt for modular story elements (like just the inciting incident or midpoint twist) — ideal for collaborative or iterative creation.
     

How to Prompt Squibler (Best Practices)

Include:

  • Genre and format (e.g., "sci-fi short film" or "episodic web series")
     
  • Clear request for structure (e.g., 3-act format, plot twist, cliffhanger)
     
  • Core story elements: character goals, settings, central conflict
     
  • Pacing expectations (short, mid-length, or feature-length content)
     

Avoid:

  • Prompts that are too vague or high-concept without context
     
  • Asking for statistical or factual content (Squibler is fiction-first)
     
  • Projects requiring visual formatting outside screenwriting (e.g., slideshow scripts)
     

Prompt Examples (2025)

Script Prompt Example:

“Write a 15-minute script for a dystopian web series pilot where a rogue AI controls a city’s water supply. Introduce the protagonist, build tension with the system, and end with a cliffhanger where they discover a hidden truth.”
 

Video Structure Prompt Example (Heist Short Film):

“Outline a 3-act structure for a 7-minute short film where a heist goes wrong.
Act 1: The setup + character motivations
Act 2: The job in progress, unexpected obstacle
Act 3: Betrayal, escape, ambiguous ending.”
 

Video Ideas Prompt Example (Cyberpunk Series):

“Generate three episodic video story ideas set in a cyberpunk city. Each should follow a young tech detective unraveling a deeper conspiracy. Include episode titles, main conflict, and teaser line for each.”
 

Series Planning Prompt (Character-Driven Drama):

“Create a high-level 5-episode arc for a drama series about a rural nurse fighting a corrupt health system. Include evolving character dynamics, main moral conflict, and how the protagonist changes over time.”

Brief Differences Overview

Squibler stands apart by focusing entirely on structured narrative development, making it ideal for screenwriters, filmmakers, and episodic content creators. It’s less about ideation and more about refining story arcs, organizing scenes, and formatting dialogue professionally. In contrast, ChatGPT is better for fast-paced creative brainstorming and tone control, while Claude focuses on emotional and ethical depth. Gemini is your best choice for creating videos with a strong visual-text connection, and Copilot leads for tutorials and code-based content. Perplexity, meanwhile, is unmatched in data-backed, research-driven scripting.


Try Squibler:

https://www.squibler.io


What ChatSonic Excels At 

ChatSonic by Writesonic is built for real-time, trend-aware content creation. It specializes in generating casual, conversational material around current events, social media trends, tech updates, and pop culture. Unlike platforms focused on storytelling or deep analysis, ChatSonic is optimized for speed, freshness, and audience relatability, making it ideal for influencers, marketers, creators, and news-style content producers.


In 2025, it remains one of the best options for producing:

  • Timely YouTube scripts and commentary videos
     
  • Short-form content based on trending hashtags
     
  • Real-time explainers and listicle-style videos
     
  • SEO-aware articles or scripts with live event relevance
     

2025 Updates: What’s New in ChatSonic

ChatSonic now integrates enhanced live web search, giving it stronger access to up-to-the-minute information across tech, entertainment, finance, and social trends. Its tone customization has also improved, letting you toggle more easily between influencer-style delivery, news brief summaries, or casual podcast-style narration. In addition, its SEO suggestion engine can now be layered directly into content generation — helpful for creators trying to stay discoverable in fast-moving niches.


How to Prompt ChatSonic (Best Practices)

To get the best results, clearly define the tone (e.g. casual, expert, energetic), the trending context, and your audience's familiarity level. ChatSonic is especially strong when you're aiming for natural dialogue, commentary, or topical engagement.


Include:

  • Current or trending topic
     
  • Casual tone instructions (e.g., “make this feel like a vlog” or “influencer style”)
     
  • Time references (e.g., “mid-2025 social media strategies”)
     
  • Real-life analogies, pop culture hooks
     

Avoid:

  • Dated language or content not grounded in current trends
     
  • Formal academic tone or deep emotional arcs
     
  • Long-term philosophical themes better suited for tools like Claude or ChatGPT
     


Prompt Examples (2025)


Conversational Script Prompt Example:
“Write a 5-minute YouTube script in a casual, vlog-style tone discussing the top 3 AI tools creators are using in 2025. Include real examples, humor, and mention of popular creator reactions.”

Video Structure Prompt Example:
“Outline a 7-minute video for a digital marketing channel. Topic: ‘Top Instagram Growth Tactics That Still Work in 2025.’ Include intro hook, audience pain points, strategy breakdowns, and call-to-action.”

Video Ideas Prompt Example:
“Give me three video ideas for a Gen Z influencer covering trending TikTok challenges and viral topics from July 2025. Include catchy titles and what makes each trend unique.”

Explainer Prompt Example:
“Create a script in the style of a fast-paced social media explainer breaking down the latest Threads vs. X platform battle. Mention user reactions, growth stats, and creator shifts.”


Differences: ChatSonic vs Other AI Platforms (Text Block)

ChatSonic is distinct from platforms like ChatGPT, Claude, or Perplexity in that it's laser-focused on real-time, trending, and conversational content. While ChatGPT is highly creative and versatile across narrative forms, and Claude excels at emotional and ethical depth, ChatSonic works best for topical scripts, live commentary, or social-first ideas. It's also more agile than Gemini when it comes to spontaneous content not tied to visuals. Unlike Copilot, which is tailored for technical clarity, or Perplexity, which prioritizes research accuracy, ChatSonic aims to help creators move quickly and stay relevant. Its sweet spot is immediacy — making content that feels like it belongs right now in a feed.


Try ChatSonic:
https://writesonic.com/chat


What TextCortex Excels At 

TextCortex is purpose-built for concise, attention-grabbing writing. In 2025, it’s a standout tool for generating short-form video scripts, social media captions, ad copy, and compact marketing messages. While other AI platforms may focus on story depth or visual integration, TextCortex thrives where brevity meets impact.


It’s ideal for:

  • Instagram Reels, TikToks, and YouTube Shorts
     
  • Ad scripts and punchy promos
     
  • Brand messaging and micro-content
     
  • Short explainers with a marketing twist
     

Think of TextCortex as the tool you reach for when you need to say a lot in a little time—with rhythm, clarity, and energy.


2025 Updates: What’s New in TextCortex


  • Smart CTA Engine: New functionality allows for optimized calls-to-action based on platform (e.g., “Link in bio” vs. “Swipe up” vs. “Shop now”).
     
  • Audience Tone Matching: Enhanced tools let you match tone to demographic segments, whether Gen Z, millennial buyers, or B2B professionals.
     
  • Creator Mode: TextCortex now includes social content packs that allow creators to generate caption + hook + script in one pass for vertical videos.
     
  • Speed Write: You can now use TextCortex to produce 10-15 second voiceover scripts in under 30 seconds, great for rapid ideation and testing content variations.


 

How to Prompt TextCortex (Best Practices)


TextCortex works best when you’re specific with format and outcome. It’s strong at trimming fat from ideas while keeping energy and clarity intact.


Include:

  • Exact video length or format (e.g., “30-second promo,” “IG Reel caption”)
     
  • A product, service, or theme
     
  • Tone (e.g., “high-energy,” “confident,” “youthful”)
     
  • Clear call to action
     

Avoid:

  • Asking for full-length scripts or slow-building narrative arcs
     
  • Vague prompts without a clear marketing angle
     
  • Long-form educational or fictional content (better suited to ChatGPT or Claude)
     

Prompt Examples (2025)


Social Script Prompt Example:
“Write a 30-second Instagram Reel script promoting a refillable water bottle. Use upbeat, eco-conscious language and end with a CTA to ‘shop the link in bio.’”


Video Structure Prompt Example:
“Outline a 3-part short-form video to promote a new digital course for freelancers.
Part 1: Grab attention with a pain point
Part 2: Quick course benefits
Part 3: Strong call to action with urgency.”


Video Ideas Prompt Example:
“Suggest three punchy Instagram video ideas for a skincare brand launching a plant-based moisturizer. Each idea should include a hook, visual concept, and caption-style CTA.”


Short Explainer Prompt Example:
“Create a script for a 20-second TikTok explaining why compostable packaging is better for the planet. Make it fast, fun, and easy to follow.”



Differences: TextCortex vs ChatGPT, Gemini, Claude, and Others (Text Block)


TextCortex is uniquely positioned for high-impact, short-form content, particularly when speed, clarity, and platform awareness matter. It doesn’t attempt to write novels or screenplays—instead, it helps brands and creators communicate quickly and persuasively in spaces like social media, email, and product videos. Compared to ChatGPT, which can generate rich, evolving narratives, TextCortex trims the storytelling down to essentials. Gemini’s strength lies in visual-text synthesis, while Claude favors deep emotional and ethical structure. Perplexity focuses on data-driven analysis, and Copilot dominates in instructional or technical guidance. TextCortex, by contrast, is the go-to for fast-paced, sales-ready language—ideal for modern content marketing.


Try TextCortex:
https://textcortex.com


PROMPTS: AI Image Generation Tools

click the drop down arrows to revel more

What Midjourney Excels At 

Midjourney has become the go-to platform for stylized, cinematic, and emotionally rich visuals. With the release of Midjourney V7, the model has evolved into a powerhouse for generating photorealistic, painterly, surreal, and concept-driven artwork. Whether you’re building a fantasy world, a product concept, or a moody cinematic shot, Midjourney brings ambiance, texture, and imagination to life with stunning visual cohesion.


It remains ideal for:

  • Cinematic scenes and concept art
     
  • Fantasy, surrealism, and dreamlike imagery
     
  • Visual storytelling, album art, fashion editorial looks
     
  • Stylized portraits and mood-heavy environments


 

2025 Midjourney V7 Updates

With V7, Midjourney has become:

  • More coherent with human anatomy and realism in characters and faces
     
  • Better at environmental consistency, light behavior, and scene depth
     
  • Improved with prompt accuracy, following complex multi-layered requests more reliably
     
  • Now supports image prompting + style reference blending, allowing creators to anchor images with more control
     

New V7 features also include:

  • Omni-Reference: Upload a reference image of a person for better character consistency across image generation
     
  • Aspect Ratio Flexibility: Create visuals in wide cinematic (e.g. 2:1) or tall social formats (e.g. 9:16)
     
  • Better handling of text in image (though still not perfect — use sparingly or for mockups only)
     

How to Prompt Midjourney (Best Practices)


To get powerful results with Midjourney, combine rich scene description with clear visual language. Think of it like writing for a production designer or concept artist.

Prompting Formula:


[Subject] + [Scene/setting] + [Mood] + [Style/art medium] + [Lighting] + [Color palette] + [Camera or composition notes] (optional)
 

Include:

  • Clear subject focus: the main character, object, or focal point
     
  • Atmosphere: moody, ethereal, dystopian, whimsical, futuristic, etc.
     
  • Style descriptors: watercolor, oil painting, cinematic still, vaporwave, brutalist, editorial fashion, etc.
     
  • Lighting conditions: soft ambient light, golden hour, harsh shadows, overcast
     
  • Color palettes: muted tones, neon gradients, warm earth tones, etc.
     
  • Optional: camera type or lens (e.g. “shot on 85mm,” “drone view,” “macro lens”)
     

Avoid:

  • Overly technical instructions or logic-based prompts (Midjourney is artistic, not analytical)
     
  • Long paragraphs — keep your prompt compact and layered
     
  • Using prompt language better suited for 3D modeling or photoreal CAD tools (Midjourney is not for precision architecture or product blueprints)


 

Prompt Examples


Prompt Example 1:
A lone cowboy walking through a foggy desert valley at dawn, cinematic mood, soft ambient light, warm earth tones, inspired by Roger Deakins, wide shot, 2.35:1 aspect ratio, hyper-detailed grainy film still


Prompt Example 2:
A whimsical forest tea party with foxes and raccoons wearing tiny hats, surreal and painterly style, inspired by Studio Ghibli and Beatrix Potter, golden hour light, soft pastels and mossy greens, magical ambiance


Prompt Example 3:
A futuristic cyberpunk alley at night, glowing neon signage, rain-slick streets, deep contrast, ultra-wide shot, 80s retrofuturism style, vivid purples and electric blues, detailed reflections, gritty urban textures


Prompt Example 4:
A fashion editorial shoot in a post-apocalyptic greenhouse, model in layered fabrics, overgrown vines, soft backlighting, moss-covered stone floor, high-fashion meets ruinpunk aesthetic, moody and textured, Vogue-style composition


Differences from Other Platforms


Midjourney remains the most visually poetic AI compared to more function-focused platforms. While Gemini blends visuals with structured logic and ChatGPT focuses on text-based prompting, Midjourney interprets mood and style more intuitively—often feeling like you're art-directing a dream. It doesn’t require code, datasets, or realism-by-default like tools such as DALL·E for prototyping or design. Instead, it leans into atmosphere, surrealism, and stylistic richness, making it a top pick for concept art, storytelling visuals, or expressive branding.


Try Midjourney:

https://www.midjourney.com


What DALL·E Excels At

DALL·E is OpenAI’s image generation model designed to turn natural language prompts into detailed, coherent, and often photorealistic or illustrative visuals. With its 2025 improvements, DALL·E now delivers stronger image consistency, more accurate compositions, and style-guided generation using image references or inpainting.


It excels in:

  • Product mockups, branding visuals, and concept design
     
  • Photo-real environments and editorial-style portraiture
     
  • Surreal and imaginative art with visual logic
     
  • Illustrations for books, learning materials, or creative decks
     

It strikes a balance between creative surrealism and design-ready realism, making it especially powerful for marketing, storytelling, teaching, and prototyping.


2025 Updates and Key Features

  • Style Transfer from Reference Images: You can now guide DALL·E to match the style of an uploaded image (e.g., a specific art style, photography look, or design palette).
     
  • Inpainting and Outpainting Tools: Easily edit, expand, or fine-tune specific sections of an image — great for adding characters, changing objects, or refining context.
     
  • Image Consistency Features (available via ChatGPT Pro): Create the same character, object, or scene in different poses or variations through “generative continuity” tools.
     
  • Integrated in ChatGPT 4o (Pro): DALL·E is now natively embedded into the ChatGPT workflow for iterative prompting, captioning, or editing support.



How to Prompt DALL·E (Best Practices)


DALL·E responds best to prompts that are:

  • Direct and specific about the subject and context
     
  • Concrete about visual and stylistic details
     
  • Balanced in length: 1–3 descriptive sentences often perform better than long paragraphs
     

Prompting Formula:

[Primary subject] + [Scene or setting] + [Style or medium] + [Lighting or mood] + [Color palette or composition details]
 

Include:

  • Subject and environment (e.g., “A girl reading a book in a rainy café”)
     
  • Visual clarity (e.g., “glass windows fogged up from the rain”)
     
  • Stylistic cues (e.g., “illustrated in watercolor style” or “shot with a soft-focus lens”)
     
  • Framing (e.g., “top-down view,” “wide angle,” “close-up portrait”)
     
  • Time of day or lighting (e.g., “sunset glow,” “soft ambient lighting”)
     

Avoid:

  • Vague prompts with no clear subject (“something dreamy” or “a nice scene”)
     
  • Overloaded instructions with conflicting styles (e.g., “photo-realistic cartoon oil painting”)
     
  • Technical jargon not related to visual attributes
     

Prompt Examples 


Prompt Example 1:
A cozy Scandinavian-style kitchen interior with natural wood cabinets, potted herbs on the windowsill, and sunlight pouring through large glass windows. Clean and minimalistic, warm earth tones, highly realistic style.


Prompt Example 2:
A futuristic cityscape at dusk, neon lights reflecting off wet streets, flying cars above, and citizens in reflective clothing walking through the plaza. Cyberpunk style, dramatic lighting, wide cinematic frame.


Prompt Example 3:
A hand-drawn children’s book illustration of a fox and a squirrel having a picnic in a flower-filled meadow, pastel color palette, friendly expressions, illustrated in watercolor style.


Prompt Example 4:
A sleek, modern electric car parked on a mountain overlook road, mist surrounding the valley below. Rendered in high-resolution, realistic lighting with a soft-focus background and glossy car surface.



Differences from Other AI Image Tools 


DALL·E differs from other image platforms like Midjourney or Stable Diffusion in its focus on prompt accuracy, realism, and structured composition. While Midjourney excels at artistic surrealism and painterly ambiance, DALL·E is more flexible across styles and often better at following logical or branded layouts — making it excellent for marketing visuals, book illustrations, and product design. It’s also deeply integrated into the ChatGPT environment, allowing users to generate, refine, and edit images directly within conversations. Compared to Gemini, which fuses text with visual elements more educationally, DALL·E remains a pure visual-generation engine tuned for practical creativity and design precision.


Try DALL·E:

https://openai.com/dall-e


What Gemini Excels At 

Gemini (formerly Bard), now part of Google’s broader multimodal ecosystem, integrates tightly with Google's Imagen model for image generation. In 2025, it excels at producing narrative-aligned, context-aware, and highly coherent visuals—especially when paired with a story, mood, or functional intent.


Gemini stands out for its ability to:

  • Bridge narrative storytelling with visual composition
     
  • Maintain logical, educational, or contextual relevance in scenes
     
  • Create visuals that feel like stills from a story, a product explainer, or an editorial
     

This makes it a strong choice for:

  • Educational content and infographics
     
  • Scene-building for storytelling or explainer videos
     
  • Brand identity visuals with clear intent
     
  • Content aligned with Google Slides, Docs, or digital classrooms
     

Imagen

As of 2025, Gemini uses Google’s Imagen 2 for visual generation within the Gemini interface. Imagen is known for generating realistic and richly detailed images, particularly when paired with structured or narrative input.

Imagen is optimized for:

  • Human-centric scenes
     
  • Professional, clean compositions
     
  • Highly detailed visual logic (great for architectural, scientific, or educational applications)
     

2025 Feature Updates

  • Visual + Text Co-generation: Gemini now allows dual prompting, where visuals are shaped by written context or story frameworks.
     
  • Style and Theme Matching: You can now ask Gemini to match brand tone, illustration style, or Google Workspace design patterns.
     
  • Narrative Progression Support: Gemini is uniquely good at generating multiple related images in a sequence (e.g., a three-panel explainer, or visual steps in a process).
     
  • Fact-anchored Visuals: Ideal for use in education or data visualization—images stay grounded in factual descriptions.
     

How to Prompt Gemini (Best Practices)

Gemini responds best when your prompt includes:

  • A narrative or functional purpose (“This image supports a lesson on coral reefs”)
     
  • Scene or environmental detail (“Include glass domes and marine life”)
     
  • Mood or visual tone (“Use ambient lighting and evoke wonder”)
     
  • Stylistic clarity (“In a warm, painterly style” or “flat, infographic-style art”)
     

Prompting Formula:

[Narrative or theme] + [Visual subject or environment] + [Tone or purpose] + [Style or medium]
 

Include:

  • A clear context or use case for the image (e.g., education, narrative, branding)
     
  • Scene layout, characters, setting, color palette
     
  • Emotional or thematic tones (e.g., awe, serenity, tension, energy)
     
  • Specify illustration vs photorealism if needed
     

Avoid:

  • Prompts lacking intent or context (“make something pretty”)
     
  • Mixing conflicting visual styles (e.g., “flat vector photorealistic steampunk painting”)
     
  • Purely abstract prompts with no usable detail
     

Prompt Examples


Prompt Example 1:
Create an image of a bustling 1920s jazz club. Show a band onstage with brass instruments, couples dancing in formal wear, and patrons at candlelit tables. Use sepia tones and art deco style to reflect the era’s mood.


Prompt Example 2:
Generate a visual of an underwater research base. Include glass observation domes, colorful coral reefs, bioluminescent sea creatures, and ambient blue-green lighting. The scene should evoke scientific curiosity and exploration.


Prompt Example 3:
Illustrate a three-step process for composting at home, in a clean flat-vector style suitable for a classroom poster. Include labeled bins, food scraps, and worms, with each step numbered.


Prompt Example 4:
Design a futuristic classroom setting for a tech-focused learning environment. Include digital blackboards, students using AR glasses, and a clean, minimalist architectural style. Bright natural lighting and diverse student characters.


Differences from Other AI Image Tools 


Gemini (with Imagen) differs from models like Midjourney or DALL·E by prioritizing narrative alignment, educational clarity, and contextual realism. While Midjourney leans into moody and artistic abstraction, and DALL·E offers flexibility between surreal and practical design, Gemini excels when your image needs to support a structured idea, learning objective, or clean brand tone. Its integration with Google’s tools also makes it particularly strong for professional or academic content. Compared to Meta AI, which focuses more on emotional and social visual storytelling, Gemini’s strength lies in intentional, clear, and logically grounded visual outputs.


Try Gemini (with Imagen):

https://gemini.google.com/


What Adobe Firefly Excels At

Adobe Firefly, part of Adobe’s advanced generative suite, is designed for creators and video producers looking to easily integrate AI-generated visuals into their projects. In 2025, it excels at producing high-quality, context-driven images that can be seamlessly used in video creation, offering a new level of creative flexibility for video editors, animators, and content creators.

Firefly shines for its ability to:

  • Generate visually striking assets that fit within specific video themes, narratives, or brand guidelines.
     
  • Deliver detailed, highly customizable images that can be further refined for use in motion graphics or video sequences.
     
  • Adapt to various video formats, from social media clips to full-length explainer videos.
     

This makes it an excellent tool for:

  • Creating stunning backgrounds, title cards, and scene elements for videos.
     
  • Enhancing video production with assets like characters, props, and environments.
     
  • Supporting video content with tailored visuals for educational, marketing, or brand purposes.
     
  • Quick and efficient content creation for video, minimizing the need for manual design work.
     

Adobe Firefly

As of 2025, Adobe Firefly is fully integrated with Adobe’s suite of creative tools, offering an AI-powered design system that prioritizes artistic control and customization. With its ability to generate professional-grade visuals, Firefly is uniquely positioned to create assets for videos, including backgrounds, animated assets, product visuals, and scene-building elements.

Firefly is optimized for:

  • Graphic assets for videos—characters, backgrounds, and objects.
     
  • Story-driven visuals that align with the narrative or tone of a video.
     
  • Smooth integration with Adobe tools like Premiere Pro, After Effects, and Photoshop, allowing easy customization of generated images for video projects.
     

2025 Feature Updates

Video Asset Creation: Firefly now includes video-oriented tools that allow users to generate images specifically for video purposes. These include background environments, motion design assets, and character elements that fit video dimensions and formats.

Dynamic Scene Adaptation: Firefly can generate visuals based on the movement or progression of a video scene. This is ideal for creating backgrounds or environments that evolve with the flow of a video.

Real-Time Customization: A new feature in Firefly allows for the instant adjustment of visual elements, enabling video editors to tweak generated images without leaving the Adobe platform. This streamlines workflows, especially for video editors working on tight deadlines.

Asset-to-Scene Matching: For users building out a full scene for video production, Firefly can now match visual elements across various assets. This ensures that all generated imagery—whether it’s a character, prop, or background—matches the same style and tone throughout the video.

How to Prompt Adobe Firefly (Best Practices)

When using Firefly for video projects, it’s essential to provide clear prompts that align with the overall tone, theme, and progression of your video. The more specific your prompt, the more tailored your assets will be, helping you seamlessly integrate them into your final product.

Firefly responds best when your prompt includes:

  • A video-related use case (“This background will be used for a science documentary”).
     
  • Detailed visual elements that match your video scenes or motion graphics (“Generate a futuristic cityscape with glowing skyscrapers for a sci-fi intro”).
     
  • Any color schemes, lighting, or style elements you’d like to match (“Dim, cold lighting with a neon color palette to evoke a dystopian feel”).
     
  • If you need the image for a dynamic video scene, mention any movement or transitions (“Create a sunset beach background that transitions into a twilight sky as the scene progresses”).
     

Prompting Formula:
[Use case] + [Visual subject or asset type] + [Style or tone] + [Color, lighting, or thematic elements]

Include:

  • The intended video scene or moment (e.g., title card, explainer, background animation).
     
  • Emotional tone or atmosphere (e.g., tense, peaceful, energetic).
     
  • Any specifics about movement or evolution in the scene (e.g., sunrise changing to midday, or a city skyline turning into a neon wonderland).
     

Avoid:

  • Generic or overly broad prompts (“Create something cool for a video”).
     
  • Mixed visual styles that might clash when animated (“Make a 1920s vintage style background with futuristic elements”).
     
  • Imbalanced prompts with conflicting moods or concepts that could be hard to visualize.
     

Prompt Examples


Prompt Example 1:
Generate a dystopian cityscape for an animated intro video. The scene should feature towering, sleek skyscrapers with glowing neon signs and empty streets. Use deep blues and purples with cold lighting to create a futuristic yet lonely atmosphere. The image will fade into a bright, bustling city center as the scene progresses.

Prompt Example 2:
Create a serene, forest-based background for an educational video on wildlife. Include tall, green trees, lush undergrowth, and soft sunlight filtering through the branches. The scene should feel calming and natural, with warm golden tones and soft green hues, evoking a sense of peace and harmony with nature.

Prompt Example 3:
Generate an abstract representation of the internet for a digital marketing video. The image should feature glowing data streams, interconnected nodes, and a glowing, circuit-like background. The style should be clean and modern, with cool blue tones and sleek, high-tech lines to convey digital connectivity.

Prompt Example 4:
Create a dynamic background for a fitness tutorial video. The scene should include a spacious gym with modern workout equipment, mirrored walls, and motivational banners. Bright, energetic lighting and vivid colors (blues, oranges, yellows) should create an atmosphere of motivation and energy.


Differences from Other AI Image Tools

Adobe Firefly distinguishes itself from other generative AI tools like DALL·E and Midjourney by focusing on user control, seamless integration with Adobe’s creative suite, and video-specific features. While DALL·E offers flexibility for a wide variety of visual styles, and Midjourney excels at abstract and artistic concepts, Firefly’s strength lies in creating assets that can be directly applied to professional video projects. It is especially powerful for designers and video editors looking for custom assets that align perfectly with their branding, storytelling, and video production needs.

Additionally, Firefly’s integration with Adobe’s other tools means that video creators can easily refine, edit, and animate AI-generated visuals within familiar programs like Premiere Pro or After Effects, offering an intuitive, streamlined workflow.


Try Adobe Firefly:

https://adobe.com/firefly


What Runway ML Excels At


Runway ML, now with its advanced Gen 4 model released in 2025, is designed to allow creators to seamlessly transition from image generation to video production—all within one platform. It excels at generating stunning, high-quality visuals that can then be transformed into dynamic videos, making it an essential tool for video creators, digital artists, and designers.


Runway ML is particularly strong in:

  • Producing contextually rich, high-resolution images tailored to specific creative projects.
     
  • Creating images with the intention of transitioning them into video, making it ideal for storytelling, visual effects, and animation.
     
  • Integrating reference images (such as photos of people, scenes, or objects) into your prompts to enhance accuracy and visual alignment with your project.
     

This makes it a top choice for:

  • Image generation that will be used as video assets or sequences.
     
  • Generating consistent visual themes or character designs that will evolve over time in video.
     
  • Quick prototyping of video content by using image assets as the building blocks for motion design.
     
  • Integrating reference photos or visual styles to keep video production coherent and on-brand.
     

Runway ML (Gen 4)

Runway ML Gen 4 takes image generation to the next level by not only creating high-quality visuals but also enabling creators to smoothly turn those images into video clips, animations, or even complex scenes. With its powerful AI model, Runway ML offers advanced capabilities for video creators who need precise images for use in their video projects, all while keeping the flow consistent across both formats.


Runway ML is optimized for:

  • Image-to-video workflows: Generate static visuals that can then be animated or turned into videos seamlessly within the Runway platform.
     
  • Reference image integration: Use up to 3 reference images to guide the AI in generating visuals that match real-world or pre-established designs.

     

2025 Feature Updates


Image-First to Video Workflow: Gen 4 introduces an enhanced pipeline where creators can generate images first and easily transform them into dynamic video content without leaving the platform. This is ideal for filmmakers, animators, and content creators looking to maintain consistent visual quality across both image and video formats.

Reference Image Upload: You can now upload up to 3 reference images to guide the AI in generating visuals that are closely aligned with real-world images or pre-existing concepts. For example, if you're creating a character, you can upload photos of a person or a specific scene to ensure the AI’s output closely matches the desired look.

Enhanced Styling Options: With Gen 4, you can now provide even more detailed stylistic guidance for your images, helping to generate visuals that fit perfectly within the overall aesthetic or tone of your video project. Whether you're working on a cinematic video or an abstract animation, Runway ML now offers enhanced flexibility for controlling the look of your visuals.


How to Prompt Runway ML (Best Practices)

When generating images for video production in Runway ML, it's crucial to be specific with your prompts to ensure that the AI produces visuals that will fit well within the larger context of your video. If you plan to use reference images, make sure to include them in your prompt to guide the AI toward the desired outcome.

Runway ML responds best when your prompt includes:

  • A clear visual intent: Describe the visual theme or purpose of the image you want to generate (“Create an atmospheric forest background for a fantasy adventure video”).
     
  • Reference images: Upload up to 3 reference images to guide the AI in generating realistic or closely matched visual elements. For example, if you're designing a character, you can upload photos of the person whose likeness you wish to replicate.
     
  • Stylistic direction: Include details about the desired mood, lighting, or artistic style (“Create a dark, moody atmosphere with soft light filtering through the trees”).
     
  • Context for video transition: If the image is meant to be turned into a video, consider how the scene or character should evolve
     

Prompting Formula:
[Visual subject or asset] + [Detailed environment or setting] + [Style, lighting, and tone] + [Reference images] + [Context for video transition]

Include:

  • A clear description of the visual goal (e.g., scene, character, object).
     
  • Environmental details like setting, background, or weather (e.g., a bustling city, a quiet library).
     
  • Lighting, color palette, and any specific aesthetic or mood (e.g., soft lighting, neon colors, vintage sepia).
     
  • One to three reference images for guiding the AI (e.g., photos of a character or similar visuals). You can reference these characters with "@" and selecting your reference image - so you can conversationally include the reference image, example: "@Reference sits on @reference2 chair in the middle of the room".
     

Avoid:

  • Vague or broad prompts that lack direction (“Make a cool image for my video”).
     
  • Mixing too many conflicting visual styles (“A futuristic city with medieval elements”).
     
  • Undefined prompts with no clear reference or context (e.g., “Create something beautiful for my video” without further detail).
     

Prompt Examples


Prompt Example 1:
Generate a futuristic character for a sci-fi video. The character should have metallic armor with glowing blue accents and a cyberpunk-style helmet. Upload reference images of a person in a sleek suit and futuristic cityscapes to guide the design. The character will be used in multiple video scenes, so make sure the design is clean and visually striking, with a strong emphasis on metallic textures and vibrant lighting.

Prompt Example 2:
Create a serene mountain landscape at sunrise for a travel video. The scene should feature snow-capped peaks, pine trees, and a soft pink glow in the sky. Use reference images of a real mountain range to guide the natural feel of the landscape. 

Prompt Example 3:
Generate an abstract neon-lit cityscape for a high-energy music video. The scene should include glowing signs, dark alleys, and illuminated streets. Use reference images of urban street scenes with neon lights to guide the visual design. 

Prompt Example 4:
Create a vintage 1920s jazz club scene for a historical documentary. The background should feature velvet curtains, art deco furniture, and a live jazz band performing on stage. Upload reference images of a 1920s jazz club for stylistic accuracy. The scene should feel nostalgic and sophisticated, suitable for a calm, story-driven video.


Differences from Other AI Image Tools

Runway ML Gen 4 differs from other AI tools like Midjourney or DALL·E by focusing specifically on the needs of video creators who require images that can be directly transitioned into motion graphics or full video sequences. While DALL·E and Midjourney offer powerful image generation, Runway ML is unique in its ability to handle both the image-first creation and seamless video integration all in one platform. Its reference image feature also sets it apart, allowing for a more guided and accurate generation of visuals—particularly when you need precise control over likenesses or stylistic alignment.

Compared to other platforms that focus more on static imagery, Runway ML is designed for video-first creators, making it ideal for video content generation where you need consistency across still images and motion elements.


Try Runway ML:

https://runwayml.com


What Meta AI Excels At

Meta AI (formerly known as part of the Meta GenAI suite) specializes in generating emotionally resonant, socially expressive, and lifestyle-driven visuals. Designed to enhance creative output across social platforms like Instagram, Facebook, Horizon, and Reels, it emphasizes people, emotion, and community energy in its generated outputs.


In 2025, Meta AI has advanced in:


  • Multi-character interactions, gestures, and body language
     
  • Social context awareness, including culturally relevant and emotionally expressive scenes
     
  • AR/VR visual integration, supporting immersive use in Meta’s mixed-reality tools
     
  • Platform-specific framing — optimized for vertical video, profile imagery, or interactive stories
     

Ideal for:

  • Social storytelling and brand content
     
  • Real-life lifestyle visuals or stylized snapshots
     
  • Community, event, or relationship-based scenes
     
  • Visuals intended for AR filters, avatars, or immersive experiences
     


2025 Updates and Key Features


  • Emotionally Intelligent Prompting: Better interpretation of feelings, moods, and social interactions
     
  • Scene Cohesion: Improved hand, face, and posture rendering — ideal for multi-person compositions
     
  • Platform Framing Awareness: Generates visuals tuned for story formats (9:16), feed posts (1:1), or immersive backdrops (panoramic/360)
     
  • Cross-Character Interaction: Can depict multiple people interacting authentically (e.g., laughter, hugging, celebrating)
     


How to Prompt Meta AI (Best Practices)


Meta AI performs best when prompts emphasize connection, atmosphere, and emotion. Think in terms of moments, not just objects or scenery.

Prompting Formula:

[Scenario or group dynamic] + [Location or environment] + [Emotional tone or action] + [Lighting or stylistic detail]
 

Include:

  • Social context: Are they friends, family, teammates, strangers?
     
  • Emotional cues: Laughter, tension, celebration, intimacy
     
  • Gestures or interactions: Holding hands, dancing, playing, arguing, etc.
     
  • Lighting and mood: Golden hour, moody evening, festive glow
     
  • Diverse representation: Age, culture, setting, body types — Meta AI handles these well
     

Avoid:

  • Static, isolated figures without emotional framing
     
  • Abstract styles or hyper-surreal concepts (better suited for Midjourney)
     
  • Technical design prompts (like CAD-style architecture or blueprint visuals)
     

Prompt Examples 


Prompt Example 1:
A group of friends gathered around a beach bonfire at sunset, laughing and roasting marshmallows, warm lighting from the flames casting soft glows on their faces. Emphasize connection and relaxed joy.


Prompt Example 2:
A lively block party on a summer evening, with neighbors dancing, kids running with sparklers, and string lights overhead. The mood is festive, inclusive, and full of motion and color.


Prompt Example 3:
Two young women hugging tightly at an airport arrival gate, tears in their eyes, surrounded by other travelers and luggage. Emphasize emotion, reunion, and natural lighting.


Prompt Example 4:
A multigenerational family cooking together in a cozy kitchen during the holidays. A child helps decorate cookies while grandparents laugh in the background. The image should feel heartwarming, candid, and detailed.



Differences from Other AI Image Platforms 


Meta AI stands apart from platforms like Midjourney and DALL·E through its focus on emotional realism and human connection. While Midjourney thrives in stylized dreamscapes and DALL·E excels at product realism or abstract design, Meta AI prioritizes natural social moments, multi-person dynamics, and lifestyle imagery that feels “captured,” not staged. It’s especially effective for creators and brands looking to tell relatable stories visually, and for building assets meant to live on social platforms. Compared to Gemini or Perplexity, Meta AI leans less into analytical precision and more into relatable, visual storytelling grounded in human emotion.


Try Meta AI:

https://ai.facebook.com


What Grok Excels At


Grok, X's generative AI tool, is tailored for creating high-quality images that can seamlessly transform into video assets within the same platform. It excels at producing visually rich, context-aware images that serve as both standalone pieces and building blocks for video content. With its deep integration into X's social media ecosystem, Grok is a powerful tool for creators who want to produce both static and dynamic visuals with minimal friction.


Grok stands out for its ability to:

  • Transform images into video assets: Generate images that can easily be used as video stills, backgrounds, or animation frames.
     
  • Contextual and trend-sensitive visuals: Produce content that aligns with current social media trends, making it perfect for creators looking to tap into the latest viral topics.
     
  • High-quality, on-brand imagery: Deliver images that match your brand’s style, tone, and message, essential for consistent content creation across video and social media platforms.
     

This makes it ideal for:

  • Video content creation: Build entire scenes and assets for video projects, including backgrounds, characters, and props, all within Grok.
     
  • Social media visual assets: Create images that align with ongoing trends and are optimized for platforms like X, making it easy to produce promotional or engaging content.
     
  • Brand visual consistency: Generate images that match your established brand identity and visual language, and quickly transition them into video formats.
     

Grok (from X)

Grok leverages X's powerful AI and its vast database of social media and cultural insights, creating images that resonate with your audience and align with current trends. As of 2025, Grok is optimized not only for image creation but for quick conversion into video elements, which is a huge advantage for content creators working in fast-paced environments.

Grok is optimized for:

  • Image-to-Video Workflow: Grok allows you to generate images specifically intended to be used in video content. Whether it's a background, a character design, or a motion graphic element, Grok helps bridge the gap from still image to dynamic video.
     
  • Real-time trend integration: Create images based on current events, trending topics, or viral aesthetics, perfect for social media-driven video content.
     
  • Reference image input: Upload images of people, settings, or objects to ensure visual consistency, especially when dealing with characters or specific objects for video sequences.
     

2025 Feature Updates

Image + Video Workflow: In 2025, Grok’s integration into X’s ecosystem now allows creators to generate images for specific video use cases. Once the image is created, you can easily transition it into video, maintaining a consistent look and feel for both static and motion content.

Reference Image Upload: A powerful feature of Grok is the ability to upload reference images. This helps the AI understand and replicate specific visual details, such as the look of a person, product, or environment. For video creators, this is especially useful when designing characters or scenes that need to match real-world references or previous design work.

Brand and Style Matching: Grok now allows users to specify not just general visual tone, but also exact brand guidelines (e.g., color palette, font types, mood). This feature is particularly useful for video creators looking to maintain a consistent brand image across both still and motion content.

Fast Content Generation: Grok is optimized to produce visuals in seconds, allowing content creators to keep up with the fast pace of social media, especially when creating real-time video content based on evolving trends.

How to Prompt Grok (Best Practices)

When using Grok for image creation, especially for video production, it’s important to craft clear and contextually relevant prompts. The more specific you are about how the image will be used in video, the better the result will be.

Grok responds best when your prompt includes:

  • Purpose for video: Specify how the image will be used in the video, e.g., as a background, a character design, or part of a scene.
     
  • Reference images: Upload reference images to guide Grok in replicating specific people, objects, or environments.
     
  • Scene or composition details: Describe the setting, lighting, and mood of the scene, especially if it will be used in a video sequence.
     
  • Video transitions: If you plan to animate or transition the image in the video, mention how it should evolve or interact with other elements in the video.
     

Prompting Formula:
[Purpose for video use] + [Visual subject or scene] + [Details for composition and style] + [Reference images] + [Video transition or animation cues]

Include:

  • Context for video use (e.g., background, character, visual effect).
     
  • Reference images for visual consistency (e.g., people, props, locations).
     
  • Scene layout, colors, and lighting preferences.
     
  • Details for how the image should work in motion (e.g., animated background, moving characters).
     

Avoid:

  • Vague or overly general prompts like “make something cool” or “create a pretty picture.”
     
  • Requests with conflicting styles or visual elements (e.g., “a Victorian building in a cyberpunk city”).
     
  • Ambiguous prompts with no clear context for the video (e.g., “just make something interesting” without any video-specific instructions).
     

Prompt Examples

Prompt Example 1:
Create an image of a futuristic city skyline for a sci-fi video. The skyline should feature towering skyscrapers, neon lights, and flying cars. Upload reference images of modern cityscapes and futuristic architecture. This image will serve as the background in a video sequence, so make sure it's dynamic and full of energy, ready for animation.

Prompt Example 2:
Generate an image of a young woman sitting in a coffee shop, with a laptop open in front of her. Upload reference images of a woman in casual attire and a cozy coffee shop. This image will be used as a character shot in a vlog-style video about remote work. The lighting should be warm and inviting, evoking a sense of comfort and productivity.

Prompt Example 3:
Create a dynamic action shot of a superhero flying over a cityscape. The superhero should be in a bright, colorful costume, and the city should have tall buildings with a sunset sky in the background. Upload reference images of superhero poses and city skylines. This image will serve as the key visual for an animated video intro.

Prompt Example 4:
Generate an image of a professional office space with modern furniture and a clean aesthetic. Upload reference images of minimalist office designs. This image will be used in a corporate promotional video, so make sure it conveys a sense of professionalism and efficiency.


Differences from Other AI Image Tools

Grok differs from models like DALL·E, Midjourney, or Adobe Firefly by focusing on real-time content creation and image-to-video workflows. While platforms like Midjourney excel at artistic and abstract visuals, Grok emphasizes practical image generation with an eye toward real-world applications like social media, marketing, and video production. Unlike DALL·E, which offers flexibility in terms of surreal or playful designs, Grok is tailored to creators who need timely, on-brand imagery for both static and dynamic use cases. It also has the advantage of being deeply integrated with X’s social media ecosystem, allowing for trend-sensitive, contextually relevant images.

Compared to other video-first tools, Grok offers an all-in-one solution for both image and video asset creation, making it an ideal choice for creators looking to produce visuals quickly, consistently, and on-brand.


Try Grok (from X):

https://x.com/grok


You'll see me use Whisk during the course for image generation. Not only do I use Whisk to generate images but I also use it to backwards engineer a prompt by extracting Whisks analysis of an image - you'll see this is section 10 in the VEO lectures. 


What Whisk Excels At

Whisk by Google Labs is an experimental creative platform that blends visual generation, storyboarding, and scene planning into one streamlined tool. Originally intended to assist writers and storytellers, Whisk has evolved into a visual ideation engine with strong narrative support.

Whisk is designed to:


  • Generate visual scenes tied to narrative logic
     
  • Support multi-panel or storyboard-style layouts
     
  • Combine character, location, and plot elements into cohesive visuals
     
  • Offer flexible style ranges — from sketch to cinematic rendering
     

It’s especially powerful for:

  • Film pre-visualization and content planning
     
  • Comic book or animated short development
     
  • Story-based educational or marketing content
     
  • Writers, directors, and creative producers developing pitch decks or treatments
     

2025 Feature Highlights

  • Scene-to-Image Translation: Input a scene from a script or idea draft, and Whisk renders it as a storyboard frame or mood panel
     
  • Character Consistency Tools: Whisk can maintain character appearances across multiple images
     
  • Dynamic Camera Direction Support: You can specify “wide shot,” “close-up,” or “over-the-shoulder” to shape composition
     
  • Integrated Worldbuilding Canvas: Whisk allows you to link visual scenes into a broader story world, ideal for series or film pre-production
     

How to Prompt Whisk (Best Practices)

Whisk responds best when your prompts include a narrative or cinematic context — imagine giving direction to a storyboard artist or visual concept designer.

Prompting Formula:

[Scene context or action] + [Characters and visual actions] + [Camera angle or style] + [Mood, lighting, and tone]
 

You can also add notes like:

  • “Panel 1 of 6” to establish sequence
     
  • “Keep consistent character from previous scene”
     
  • “Storyboard layout” or “cinematic frame” to guide framing
     

Include:

  • Action and character behavior ("A young girl races through a cornfield at dusk")
     
  • Cinematic tone ("dramatic backlighting, handheld camera feel")
     
  • Story moment or tension ("She looks back, someone is chasing her")
     
  • Visual direction ("Overhead shot, grainy texture, 35mm film look")
     
  • Scene tags or beats to anchor continuity ("Scene 3: The Escape")
     

Avoid:

  • Vague or emotionless concepts (“a field” or “a city”)
     
  • Mixed visual styles without direction (“sketchy, hyperreal, cartoon, cinematic”)
     
  • Single-word prompts or mood-only inputs (Whisk thrives on story + style)
     

Prompt Examples 


Prompt Example 1:
Scene 2: The Escape – A teenage girl runs through a cornfield at dusk, her face filled with fear. A shadowy figure is seen in the distance. Use a wide shot with golden backlight, grainy 35mm film style. Mood: suspenseful, urgent.


Prompt Example 2:
Scene 5 – Two astronauts float outside a damaged space station. One reaches for a broken panel, the Earth below in frame. Use a close-up on their helmet reflections, soft lighting, cinematic realism.


Prompt Example 3:
Panel 1 of 4: A wizard stands at the edge of a cliff, casting lightning toward a dragon overhead. Storm clouds churn behind him. Dramatic angle from below, painterly fantasy style, cool blue and violet tones.


Prompt Example 4:
Marketing Visual: A cozy kitchen scene with a mother and child baking cookies. Use soft warm lighting, vintage textures, and a mid-shot that shows their joyful expressions and flour-covered hands. Include kitchen props and ingredients in soft focus.


Differences from Other AI Image Tools

Whisk differs from tools like DALL·E and Midjourney by focusing on scene logic, narrative consistency, and visual continuity. It’s not just about creating a beautiful image — it’s about generating a moment that feels like part of a larger story. Compared to Gemini (which is more structured and info-aligned) and Midjourney (more abstract and artistic), Whisk is built for storytellers, giving you tools to explore, develop, and visualize entire narratives — from script to screen. It also stands apart in its support for camera direction, character continuity, and story-driven visual design.


Try Whisk (Google Labs):

https://labs.google


PROMPTS: AI Video Generation Tools

click the drop down arrows to revel more

Creating Effective Prompts for VEO 3 (Text-to-Video)

What is VEO 3?

VEO 3 is an advanced text-to-video AI tool that revolutionizes the way content creators produce video. As of 2025, it’s one of the leading platforms for transforming detailed text prompts into fully-fledged videos, including visuals, characters, dialogue, sound effects, and lip-syncing speech. Unlike traditional video production methods, VEO 3 combines the power of AI to generate entire videos from descriptive text inputs, allowing for an incredibly high degree of customization in terms of visual style, audio, and dialogue delivery.

VEO 3 excels at:

  • Text-to-Video Generation: Transforming detailed textual descriptions directly into videos, ranging from realistic cinematic shots to animated or abstract styles.
     
  • Realistic or Stylized Video: Produce videos in any desired style, whether hyper-realistic, stylized animation, or artistic abstraction.
     
  • Audio and Lip-syncing: Along with generating the visuals, VEO 3 can generate natural speech in various tones and accents, ensuring perfect lip-syncing for characters in the video.
     
  • Frame and Ingredients: It can create individual frames as well as detailed "ingredients" (assets like objects, backdrops, and characters) that can be used in different scenes of a video, giving you total control over every part of your creation.
     

What’s Needed for the Best Text-to-Video Prompt?


To get the most out of VEO 3’s text-to-video capabilities, it's essential to provide a comprehensive, clear, and detailed prompt that covers all the aspects of the video you want to create. This includes everything from visual style to character actions and dialogue. Here’s how to craft the perfect prompt.


1. Video Narrative or Purpose

Start by clearly defining the purpose or storyline of the video. Is it for an advertisement, educational content, a short film, or a social media post? This gives the AI a foundation on which to base all other elements.

Example:
"A tutorial explaining how to grow tomatoes in a backyard garden."


2. Visuals: Environment, Style, and Composition

VEO 3 excels at creating various visual styles, from realistic landscapes to abstract illustrations. Specify the type of environment, the color palette, and the visual tone of the video. If you're aiming for a certain aesthetic (like cartoonish, noir, or hyper-realistic), be explicit.

Details to Include:

  • Scene: Describe the setting in detail (e.g., a lush backyard garden with tomatoes growing on vines, an interior office space).
     
  • Lighting: Specify the type of lighting (e.g., soft sunlight, warm cozy lighting, neon glow).
     
  • Visual Style: Realistic, hand-drawn animation, flat illustration, cinematic, etc.
     
  • Camera Angles and Framing: Specify if you want a close-up, wide shot, over-the-shoulder view, or top-down shot.
     

Example:
"The scene is a bright backyard garden, with a raised wooden bed filled with tomato plants. There is soft sunlight filtering through the trees, casting a warm golden glow. The camera zooms in to focus on the tomatoes as they ripen."


3. Characters and Actions

Define who’s in the scene, their appearance, clothing, and actions. If characters are involved, mention their age, gender, ethnicity, and expressions. If you want multiple characters, detail how they should interact.

Details to Include:

  • Character Descriptions: Age, gender, clothing, posture, and any specific features.
     
  • Action/Movement: What are the characters doing? Are they interacting with objects? Moving around?
     
  • Emotional Tone: What emotions should the characters convey? Excitement, sadness, confusion, joy?
     

Example:
"A middle-aged Caucasian woman in gardening clothes is crouching beside the raised bed, inspecting the tomatoes. She’s smiling as she gently picks a ripe tomato."


4. Dialogue & Lip-syncing

VEO 3 allows you to generate dialogue that will be synced with the characters' lip movements. Specify the exact lines or speech, the tone of voice, and the mood.

Details to Include:

  • Exact Dialogue: Provide the exact words or sentences that need to be spoken.
     
  • Tone of Voice: Specify if the dialogue should be happy, serious, educational, sarcastic, etc.
     
  • Accents or Speech Styles: If you have a specific accent or speech style in mind (e.g., British accent, fast-paced speech), make sure to include that.
     
  • Voice-over or Character Interaction: If it’s a voice-over or if characters are having a conversation, mention who says what and when.
     

Example:
"The woman says, 'These tomatoes are the best I’ve ever grown. If you want to learn how to do it, keep watching!' She smiles as she looks directly into the camera."


5. Sound Effects and Music

VEO 3 can generate background sounds or music to match the mood of your video. Specify any particular sound effects you need (e.g., footsteps, bird chirping) or the type of music (e.g., upbeat, relaxing, suspenseful).

Details to Include:

  • Background Sounds: Birds chirping, rustling leaves, cars passing by, etc.
     
  • Music Style: Classical, ambient, upbeat, etc.
     
  • Sound Effects: If certain actions should have corresponding sounds (e.g., a tomato being picked with a slight crunch, the rustling of leaves).
     

Example:
"Soft birds chirping in the background, with gentle acoustic guitar music playing to set a calming, pleasant tone."


6. Audio Syncing and Timing

Make sure to define how the audio (dialogue, sounds, and music) should sync with the visual elements. Timing plays a crucial role in how natural and engaging the video feels.

Details to Include:

  • Speech Timing: When does the character speak? Ensure the dialogue aligns with the action in the scene.
     
  • Sound and Music Fade: Specify if and when background sounds or music should fade in or out.
     

Example:
"As the woman picks the tomato, she starts speaking, and the background music gradually fades in. The music then fades out slightly when she finishes speaking, allowing the sound of the garden to take over."


7. Special Effects (Optional)

If you’re going for something more dynamic or unique, mention any special effects (e.g., zoom-ins, transitions, explosions, glowing objects, or any fantastical elements).

Example:
"As the woman holds the ripe tomato, the scene should transition with a gentle glow, highlighting the fruit in her hand."


Prompt Example

Here’s a complete example of a well-rounded prompt for VEO 3:

Prompt Example:
"A close-up shot of a woman in her late 30s, wearing a light blue gardening outfit, standing in a backyard garden with raised beds full of tomatoes. The sunlight is soft and golden as she picks a ripe tomato. As she smiles and looks directly into the camera, she says, 'These tomatoes are the best I’ve ever grown. If you want to learn how to do it, keep watching!' The background is peaceful, with birds chirping, and gentle acoustic guitar music is playing softly. The camera zooms in slightly as she picks the tomato, and the sound of the tomato being picked is audible with a slight crunch."


Key Takeaways for a Full and Effective Text-to-Video Prompt

  • Narrative & Purpose: Always start with the overall story or purpose of the video.
     
  • Detailed Visuals: Provide information on the setting, characters, camera angles, and lighting.
     
  • Dialogue: Include exact dialogue with emotional tone and syncing instructions.
     
  • Audio Elements: Don’t forget background music, sound effects, and any specific audio needs.
     
  • Character Actions & Movement: Specify what the characters are doing and how their actions sync with the dialogue.
     
  • Clarity & Specificity: The more detailed and specific you are, the better VEO 3 will deliver.
     

By following this structure, you'll be able to craft detailed prompts that give VEO 3 everything it needs to create high-quality, engaging videos directly from text!


Try VEO 3

https://deepmind.google/models/veo/


Mastering SORA Prompts for AI Video Creation

SORA offers powerful tools for generating both text-to-video and image-to-video outputs. While using an image as a starting point ensures visual fidelity, crafting detailed text prompts remains crucial. The best practices for each method share similarities, with specific nuances for effective results with text to video (you will need to describe the characters, clothing and scene with more detail - like we did with AI Image generation).


Key Guidelines for SORA Prompting

1. Start with Detailed Descriptions
For text-to-video, remember to include essential visual and descriptive elements, such as:

  • Characters: Appearance, clothing, age, features, and personality.
  • Setting: The environment with intricate details, like decor or weather.
  • Mood and Style: Define the atmosphere or artistic direction.
  • Camera Movements: Specify angles or techniques (e.g., zoom-in, dolly shot).
  • Shot Types: Wide shots, close-ups, or establishing shots.


Example:
A cinematic shot of a 35-year-old knight in full metal armor scaling a snowy peak at dawn. Golden sunlight reflects off the snow, emphasizing the epic fantasy tone in a wide-angle, sweeping shot.


2. Strike a Balance Between Brevity and Detail

  • Brief Prompts: Allow SORA to interpret and add creative input (e.g., "A magical forest at twilight").
  • Detailed Prompts: Provide clear, specific guidance for precise results (e.g., "An astronaut exploring a red desert planet under a soft pink sky with a wide lens").


3. Stay Grounded in Realistic Concepts
Avoid overly abstract or impossible visuals that may confuse the AI. Use scenarios that can be rendered believably.

Avoid: "A dragon exploding into fireworks made of water."
Better: "A dragon flying over a moonlit forest, its scales glistening in silver light."


4. Utilize Emotional and Lighting Cues
Integrate emotional undertones and lighting descriptions to add depth and context. Lighting conveys mood and enhances storytelling.


Example:
A tranquil lake surrounded by glowing autumn foliage at sunrise, casting warm, golden light on the scene, evoking serenity and reflection.


5. Use Cinematic Language
Incorporate filmmaking terminology to refine the visual style and composition.

  • Camera Angles: "Over-the-shoulder," "bird’s-eye view."
  • Lens Types: "Wide-angle," "50mm prime lens."
  • Techniques: "Tracking shot," "soft focus," "dramatic backlighting."


Example:
A protagonist silhouetted against a fiery sunset, captured with a telephoto lens, as the camera slowly zooms in for dramatic tension.


6. Experiment with Artistic Styles
Define specific art or film styles to achieve distinctive aesthetics.

  • Examples: “Claymation,” “black-and-white noir,” “cyberpunk neon.”


Example:
A bustling 1920s jazz club filled with dancers, shot in black and white with a grainy film effect reminiscent of silent cinema.


For more insights, explore the styles guide at AI Video School Styles.


Runway focuses on real-time video editing and generation, offering tools that allow users to create videos with dynamic visuals and seamless transitions.

  • Overview: Runway is versatile in both text-to-video and image-to-video applications. It excels in generating motion graphics, dynamic transitions, and real-time effects.
  • Differences: It’s optimized for creating visuals that transition smoothly into video format, making it ideal for short films, commercials, and creative content with motion and effects.


Prompt Structure for Text-to-Video:

  • Start with a Clear Scene Description → Describe Key Actions and Visuals → Include Motion Cues or Effects.
  • Text-to-Video Prompt Example: “Generate a video showing a skateboarder performing tricks at an urban skate park at sunset. The camera should pan to follow the skater’s movements, with smooth transitions between shots. Add a lens flare effect as the sun sets behind the skater during a jump.”


Prompt Structure for Image-to-Video:

  • Identify the Source Image and Key Visual Elements → Describe Desired Motion or Effects → Specify Scene Transitions or Animations.
  • Image-to-Video Prompt Example: “Using an image of a neon-lit cityscape, generate a short video sequence with animated raindrops falling, reflections on wet pavement, and neon lights flickering. Add subtle camera movement to create depth and immersion.”
  • Include: Motion details, transition effects, specific camera movements.
  • Avoid: Vague descriptions of effects or movements.


  • Website: https://runwayml.com


Haiper is designed for generating short, engaging videos with a focus on storytelling and dynamic visual elements.

  • Overview: Hapier excels at producing narrative-driven videos that incorporate text, images, and video clips. It’s great for generating social media content, explainer videos, or engaging promotional videos.
  • Differences: Hapier is optimized for storyboarding and dynamic transitions, making it suitable for concise storytelling with clear visual and narrative arcs.


Prompt Structure for Text-to-Video:

  • Define the Main Theme or Story → Provide Visual and Narrative Cues → Mention Key Transitions and Text Overlays.
  • Text-to-Video Prompt Example: “Create a 30-second promotional video about eco-friendly products. Start with shots of nature, like forests and oceans, and include on-screen text highlighting sustainability benefits. Transition to close-ups of products, with upbeat background music and quick cuts to maintain energy.”


Prompt Structure for Image-to-Video:

  • Describe the Visual Style and Key Elements of the Source Image → Include Details for Animations or Motion Graphics → Specify Narrative or Text Overlays.
  • Image-to-Video Prompt Example: “Using an image of a modern workspace, create a short video that animates text describing productivity tips. The camera should gently zoom in and out, while icons and arrows highlight key features in the workspace.”
  • Include: Clear transitions, text overlays, engaging cuts.
  • Avoid: Complex effects without clear narrative guidance.


  • Website: https://haiper.ai/


Pika focuses on generating engaging, visually appealing videos with a simple and intuitive interface, making it ideal for marketing and social media videos.

  • Overview: Pika excels at creating eye-catching videos with smooth transitions, motion graphics, and dynamic animations. It’s particularly effective for quick social media content and promotional visuals.
  • Differences: Pika’s strength lies in its ease of use and accessibility, making it great for non-technical users looking to create professional-looking videos quickly.


Prompt Structure for Text-to-Video:

  • Outline the Key Message or Focus → Describe Visual and Motion Elements → Include Specific Animations or Visual Effects.
  • Text-to-Video Prompt Example: “Generate a 15-second social media ad promoting a new fitness app. Start with an animated logo reveal, followed by dynamic shots of users exercising, with animated text showing app features. End with a call-to-action and energetic background music.”


Prompt Structure for Image-to-Video:

  • Identify the Source Image and Describe the Intended Visual Style → Include Details for Motion Graphics or Effects → Specify Transitions and Final Message.
  • Image-to-Video Prompt Example: “Using an image of a tech product, create a video sequence with rotating views of the product, highlighting its features with animated text and glowing effects. Add a smooth transition to an end screen with a call-to-action.”
  • Include: Specific visual effects, product highlights, and key messaging.
  • Avoid: Overloading with too many effects or unrelated visuals.


  • Website: https://pika.art


PROMPTS: AI Audio Generation Tools

click the drop down arrows to revel more

ElevenLabs is an amazing AI tool for audio; narration and sound effects. Use it for text to speech and voiceover changer tools, note: prompting may not be needed. However, you may want to prompt for sound effects. 


ElevenLabs specializes in generating realistic voiceovers, custom sound effects, and synthetic speech with natural intonations. It excels in creating soundscapes and voice-based audio elements, making it ideal for dialogue and sound effect design.

  • Overview: ElevenLabs focuses on creating high-quality, realistic voiceovers and sound effects. It’s great for adding narrations, character voices, or environmental sounds to videos.
  • Differences: ElevenLabs is optimized for generating sound effects, voiceovers, and speech elements. It’s ideal for scenarios where specific vocal expressions or environmental sounds are needed.

Prompt Structure for Sound Effects:

  • Define the Type of Sound Effect → Describe the Context or Scene → Specify Desired Intensity or Emotion.
  • Prompt Example 1: “Generate a soft, ambient wind sound with a gentle rustling of leaves in the background. The sound should be calming and evoke the feeling of a quiet forest in the early morning.”


  • Include: Specific traits, ambient sound descriptions, context for sound effects.
  • Avoid: Vague requests without emotional or tonal guidance.
  • Website: https://try.elevenlabs.io/welj5wt1zzze


SUNO is designed for creating full audio tracks, such as music compositions, backing tracks, and complete soundscapes for video content. It’s versatile for generating anything from background scores to full-length music tracks.


  • Overview: SUNO focuses on creating complete audio compositions, making it ideal for adding thematic background music or customized soundtracks to videos. It can generate tracks based on mood, tempo, and style preferences.
  • Differences: SUNO excels at generating full audio tracks rather than individual sound effects, allowing users to set the entire tone for a video or scene including tracks with vocals and even specific lyrics or themes.


Prompt Structure for Full Audio Tracks:


  • Specify the Type of Track or Style → Describe the Desired Mood and Tempo → Include Instrument Preferences or Key Changes.
  • Prompt Example 1: “Generate a cinematic orchestral soundtrack with a rising tempo and uplifting mood. Include strings, brass, and light percussion to evoke a sense of triumph and achievement.”
  • Prompt Example 2: “Create a chill lo-fi background track for a study video. Use soft piano chords, a relaxed beat, and ambient synths to maintain a calming atmosphere.”


Prompt Structure for Background Scores:

  • Describe the Scene or Visual → Specify the Desired Emotion or Theme → Mention Instrument Preferences.
  • Background Score Prompt Example: “Generate a suspenseful background score for a mystery scene. The track should have a slow tempo, with light strings, echoing piano notes, and occasional low-frequency hums to build tension.”
  • Include: Specific music styles, desired emotions, and instrumental details.
  • Avoid: Overly broad requests without musical or thematic context.


Example: Generating a country song with a specific theme:

To create a country song using SUNO, you’ll need to provide clear details about the theme, mood, and specific elements you want to include. Country songs are often narrative-driven, focusing on relatable stories, emotions, and imagery.


Key Elements to Include in the Prompt:

  1. Theme and Story: Describe the central theme or storyline of the song. Country songs frequently revolve around love, heartache, nostalgia, small-town life, or overcoming challenges.
  2. Emotion and Tone: Specify the emotional tone of the song, whether it's upbeat, melancholic, reflective, or celebratory.
  3. Instruments: Mention traditional country instruments like acoustic guitar, banjo, fiddle, harmonica, or pedal steel guitar to create an authentic sound.
  4. Pace/Tempo: Indicate the pace of the song, whether it's slow and soulful, mid-tempo, or upbeat.
  5. Lyrics Style (if relevant): Include visual or relatable imagery common to country music, like references to rural settings, nature, relationships, or familiar life experiences.


Example Prompt:

“Generate a country song with a nostalgic and reflective mood, focusing on the theme of returning home after years of being away. The song should have a slow to mid-tempo pace with acoustic guitar, soft fiddle, and light harmonica. Include lyrics that evoke imagery of dirt roads, old memories, and reconnecting with family and old friends. The tone should be warm and heartfelt, creating a sense of longing and comfort.”


Explanation of What’s Included:

  • Theme: Returning home and reconnecting with the past.
  • Emotion and Tone: Nostalgic, reflective, warm, and heartfelt.
  • Instruments: Acoustic guitar, fiddle, harmonica.
  • Pace: Slow to mid-tempo to match the emotional tone.
  • Lyrics Style: Imagery of dirt roads, old memories, and connections.


Why These Elements Are Needed:

These elements give SUNO clear guidelines on the theme and musical style, ensuring that the generated track captures the essence of a classic country song. The mention of instruments and specific imagery helps refine the sound and narrative focus, making it more relatable and emotionally resonant.


  • Website: https://suno.com/create


Subscribe

Get offers, updates, free resources - stay in the loop (no spam I promise... I hate spam!)

Go back to the course here:

  • YOUR COURSE PAGES

AI Video School

ลิขสิทธิ์ ©2025 AI Video School - สงวนสิทธิ์ทุกประการ

This website uses cookies.

We use cookies to analyze website traffic and improve the performance of our website. By accepting our use of cookies, we will aggregate your data with all other user data.

RejectAccept