ClawKit Logo
ClawKitReliability Toolkit
Back to Registry
Official Verified media Safety 4/5

veed-ugc

Generate UGC-style promotional videos with AI lip-sync. Takes an image (person with product from Morpheus/Ad-Ready) and a script (pure dialogue), creates a video of the person speaking. Uses ElevenLabs for voice synthesis.

Why use this skill?

Automate high-converting UGC video production with AI lip-sync. Convert static product photos into engaging, professional video scripts using ElevenLabs.

skill-install — Terminal

Install via CLI (Recommended)

clawhub install openclaw/skills/skills/pauldelavallaz/veed-ugc
Or

What This Skill Does

The veed-ugc skill is a specialized tool for generating high-quality User Generated Content (UGC) style promotional videos. By leveraging ComfyDeploy's sophisticated Veed-UGC workflow, this agent takes a static image—typically a person holding a product generated via Morpheus or Ad-Ready—and synchronizes it with an AI-generated audio track. It uses ElevenLabs for voice synthesis, ensuring that the character's lip movements match the spoken dialogue perfectly. This tool is designed to bridge the gap between static product photography and engaging, persuasive video marketing assets.

Installation

To integrate this skill into your environment, use the OpenClaw CLI tool. Run the following command in your terminal: clawhub install openclaw/skills/skills/pauldelavallaz/veed-ugc Ensure you have the necessary API keys configured for ComfyDeploy and ElevenLabs, as these are required for the workflow to communicate with external generative services.

Use Cases

  • E-commerce Scaling: Quickly create hundreds of variations of product testimonials for social media advertising.
  • A/B Testing: Generate multiple versions of a video with different scripts to determine which dialogue converts better for your target audience.
  • Influencer Marketing Automation: Simulate influencer-style reviews without the overhead of physical production for every iteration of a campaign.
  • Localized Content: Easily swap voice IDs to create versions of your promotional content in different dialects or languages while maintaining consistent visual branding.

Example Prompts

  1. "Create a UGC video using the product image at ./ad-ready-shot.png. Use the script: 'I honestly can't believe how much time this tool saves me every day.'"
  2. "Generate a promotional video with the image from the last Morpheus run. Use the default voice and say: 'Stop scrolling! If you haven't tried this yet, you are missing out.'"
  3. "Make a 15-second promo clip for our new coffee brand using my latest portrait. Script: 'The boldest roast I've ever tasted, and it's delivered right to my door.'"

Tips & Limitations

  • Strict Dialogue: The most critical constraint is that the script input must contain pure dialogue. Do not include stage directions, tone descriptors (e.g., [excited]), or labels like 'Scene 1'. Any extra text will be synthesized as audio and potentially cause lip-sync issues.
  • Input Quality: The final output quality heavily depends on the clarity of the initial image. Ensure the person in the image is clearly visible and the product is well-lit for the best results.
  • Voice Mapping: While the default voice is optimized for general marketing, experimenting with different ElevenLabs IDs can significantly alter the brand sentiment of your video.

Metadata

Stars1217
Views1
Updated2026-02-20
View Author Profile
AI Skill Finder

Not sure this is the right skill?

Describe what you want to build — we'll match you to the best skill from 16,000+ options.

Find the right skill
Add to Configuration

Paste this into your clawhub.json to enable this plugin.

{
  "plugins": {
    "official-pauldelavallaz-veed-ugc": {
      "enabled": true,
      "auto_update": true
    }
  }
}

Tags(AI)

#video-generation#ugc#lip-sync#ai-marketing#elevenlabs
Safety Score: 4/5

Flags: external-api, file-read, file-write

Related Skills

morpheus-fashion-design

Generate professional advertising images with AI models holding/wearing products. ✅ USE WHEN: - Need a person/model in the image WITH a product - Creating fashion ads, product campaigns, commercial photography - Want consistent model face across multiple shots - Need professional lighting/camera simulation - Input: product image + model reference (or catalog) ❌ DON'T USE WHEN: - Just editing/modifying an existing image → use nano-banana-pro - Product-only shot without a person → use nano-banana-pro - Already have the hero image, need variations → use multishot-ugc - Need video, not image → use veed-ugc after generating image - URL-based product fetch with brand profile → use ad-ready instead OUTPUT: Single high-quality PNG image (2K-4K resolution)

pauldelavallaz 1217

ugc-manual

Generate lip-sync video from image + user's own audio recording. ✅ USE WHEN: - User provides their OWN audio file (voice recording) - Want to sync image to specific audio/voice - User recorded the script themselves - Need exact audio timing preserved ❌ DON'T USE WHEN: - User provides text script (not audio) → use veed-ugc - Need AI to generate the voice → use veed-ugc - Don't have audio file yet → use veed-ugc with script INPUT: Image + audio file (user's recording) OUTPUT: MP4 video with lip-sync to provided audio KEY DIFFERENCE: veed-ugc = script → AI voice → video ugc-manual = user audio → video (no voice generation)

pauldelavallaz 1217

sora

Generate videos using OpenAI's Sora API. Use when the user asks to generate, create, or make videos from text prompts or reference images. Supports image-to-video generation with automatic resizing.

pauldelavallaz 1217

ad-ready

Generate advertising images automatically from a product URL + brand profile. ✅ USE WHEN: - User provides a product URL (e-commerce link) - Want automated product scraping + image generation - Have a brand profile to apply (70+ brands available) - Need funnel-stage targeting (awareness/consideration/conversion) - Want AI to auto-select model, scene, lighting based on brand ❌ DON'T USE WHEN: - User provides local product image file → use morpheus-fashion-design - Don't need a person in the image → use nano-banana-pro - Want manual control over model, scene, packs → use morpheus-fashion-design - Already have hero image, need variations → use multishot-ugc - Need video output → use veed-ugc after image generation INPUT: Product URL + brand name (optional) + funnel stage (optional) OUTPUT: PNG advertising image with product + model

pauldelavallaz 1217

sora

Generate videos from text prompts or reference images using OpenAI Sora. ✅ USE WHEN: - Need AI-generated video from text description - Want image-to-video (animate a still image) - Creating cinematic/artistic video content - Need motion/animation without lip-sync ❌ DON'T USE WHEN: - Need lip-sync (person speaking) → use veed-ugc or ugc-manual - Just need image generation → use nano-banana-pro or morpheus - Editing existing videos → use Remotion - Need UGC-style talking head → use veed-ugc INPUT: Text prompt + optional reference image OUTPUT: MP4 video (various resolutions/durations)

pauldelavallaz 1217