Claude-skill-registry higgsfield-operator
Master operator guide for Higgsfield AI video generation platform. Covers all 70+ camera presets, 23+ VFX effects, Soul ID character consistency, multi-model integration (Sora 2, Veo 3.1, WAN 2.5, Kling 2.6), Python SDK, and production workflows.
git clone https://github.com/majiayu000/claude-skill-registry
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/data/higgsfield-operator" ~/.claude/skills/majiayu000-claude-skill-registry-higgsfield-operator && rm -rf "$T"
skills/data/higgsfield-operator/SKILL.mdHiggsfield Operator
Master guide for Higgsfield AI - the multi-model video generation platform with 70+ cinematic camera presets, 23+ VFX effects, and integration with Sora 2, Google Veo 3.1, WAN 2.5, Kling 2.6, and more. This skill covers everything from basic image-to-video generation to advanced production workflows.
Platform Overview
Higgsfield aggregates 15+ premium AI video models under one platform:
- OpenAI Sora 2 / Sora 2 Pro - Text-to-video with multi-scene support
- Google Veo 3.1 - UGC Builder for talking heads
- WAN 2.5 - Audio-synced video with camera controls
- Kling 2.6 - High-fidelity video generation
- Nano Banana Pro - Fast generation (unlimited on Ultimate+)
Key Differentiator: Unlike single-model tools, Higgsfield layers professional controls (camera simulation, character consistency, lip-sync) on top of best-in-class AI models.
When to Use This Skill
Use this skill when:
- Creating AI-generated video content (social, ads, content)
- Applying cinematic camera movements to static images
- Adding VFX without green screens or post-production
- Maintaining character consistency across multiple videos (Soul ID)
- Building talking-head videos with lip-sync
- Automating video generation via Python SDK
- Choosing between Higgsfield's integrated models
Not recommended for:
- Real-time video editing (use traditional NLEs)
- Video longer than 1 minute (current AI video limits)
- Precise frame-by-frame control (AI generates autonomously)
Quick Reference
| Action | Method/Tool |
|---|---|
| Image-to-Video | Upload image + select camera preset |
| Text-to-Video | Sora 2 or WAN 2.5 with text prompt |
| Character Consistency | Soul ID (upload 10+ reference photos) |
| Talking Head | UGC Builder (Veo 3.1) + Lipsync Studio |
| VFX Application | Select effect from 23+ presets |
| API Generation | Python SDK |
Core Workflows
Workflow 1: Image-to-Video with Camera Motion
Goal: Transform a static image into a cinematic video clip
Steps:
- Upload high-quality source image (1024x1024+ recommended)
- Select camera preset from 70+ options
- Optionally stack up to 3 movements
- Choose aspect ratio (16:9, 9:16, 1:1)
- Generate and download
Camera Preset Categories:
| Category | Examples | Best For |
|---|---|---|
| Dolly | Dolly In, Dolly Out, Dolly Zoom | Product reveals, emphasis |
| Pan | Whip Pan, Pan Left/Right | Scene transitions, reveals |
| Tilt | Tilt Up/Down, Dutch Tilt | Dramatic reveals, horror |
| Tracking | Tracking Shot, Follow Shot | Action, chase scenes |
| Aerial | FPV Drone, Crane Shot, Helicopter | Establishing shots |
| Specialty | Bullet Time, 360 Rotation, Crash Zoom | Action, stylized content |
| Stabilized | Static, Locked Frame | Dialogue, interviews |
Pro Tip: Stack movements for complex motion: Crane + Dolly Zoom + Rotation = Christopher Nolan vibes.
Workflow 2: Soul ID Character Consistency
Goal: Maintain identical character appearance across multiple videos
Steps:
- Upload 10+ clear reference photos
- Different angles (front, 3/4, profile)
- Various expressions
- Consistent lighting preferred
- System creates digital twin capturing:
- Face shape and structure
- Hair style and color
- Expression patterns
- Posture characteristics
- Select from 50+ style presets (Amalfi Summer, Gorpcore Outdoor, 0.5 Selfie, etc.)
- Generate videos with consistent character
Example Prompt with Soul ID:
A young woman walks through a busy Tokyo street at night, neon lights reflecting off wet pavement. [Soul ID: @my-character] Camera: Tracking Shot
Use Cases:
- Brand ambassadors across campaign
- Character-driven content series
- UGC-style ads with consistent "creator"
- Storytelling with recurring characters
Workflow 3: VFX Effects Application
Goal: Add blockbuster VFX to videos without post-production
Complete VFX Effects List:
| Category | Effects |
|---|---|
| Explosions | Building Explosion, Clone Explosion, Head Explosion, Plasma Explosion, Car Explosion |
| Fire/Heat | Fire Element, Firelava, Firework, Flame On, Flame Transition, Fire Breath, Set On Fire |
| Transformations | Turning Metal, Cyborg, Animalization, Mystification, Gorilla Transfer, Monstrosity |
| Disintegration | Disintegration, Datamosh, Morphskin |
| Superpowers | Thunder God, Invisible, Luminous Gaze, Levitation, Hero Flight, I Can Fly |
| Nature | Earth Element, Earth Wave, Garden Bloom, Nature Bloom, Sakura Petals, Northern Lights |
| Transitions | Display Transition, Flying Cam Transition, Smoke Transition, Melt Transition, Seamless Transition |
| Character | Black Tears, Glowing Fish, Shadow Smoke, Tentacles, Symbiote, Angel Wings |
| Environmental | Aquarium, Flood, Cotton Cloud, Money Rain, Pizza Fall |
| Style | Glitch, Point Cloud, Polygon, Portal, Saint Glow, Paint Splash, Powder Explosion |
Combination Effects (Beta):
- Action Run + Set on Fire
- Building Explosion + Disintegration
- Car Chasing + Building Explosion
- Crash Zoom In + Face Punch
Application Steps:
- Upload source image or generate video
- Select VFX effect from library
- Preview and adjust intensity (if available)
- Combine with camera movement
- Generate final output
Workflow 4: Talking Head Videos (UGC Builder)
Goal: Create realistic talking-head content for ads and testimonials
Powered by: Google Veo 3.1 + Lipsync Studio
Steps:
- Upload character image or use Soul ID
- Input script text or upload audio
- Select voice (AI synthesis or voice clone)
- Apply style preset (professional, casual, energetic)
- Generate lip-synced video
Best Practices:
- Keep clips under 30 seconds for best quality
- Use clear, well-lit face images
- Script natural, conversational language
- Test multiple voice options
Workflow 5: Python SDK Integration
Goal: Automate video generation programmatically
Installation:
pip install higgsfield-client
Authentication:
# Option 1: Combined key export HF_KEY="your-api-key:your-api-secret" # Option 2: Separate keys export HF_API_KEY="your-api-key" export HF_API_SECRET="your-api-secret"
Get credentials from Higgsfield Cloud
Synchronous Generation:
import higgsfield_client as hf # Upload source image image_url = hf.upload_file("./my-image.jpg") # Generate video with camera preset result = hf.subscribe( arguments={ "image": image_url, "camera_preset": "dolly_in", "aspect_ratio": "16:9", "model": "wan_2.5" } ) print(f"Video URL: {result['output_url']}")
Asynchronous Generation with Polling:
import higgsfield_client as hf # Submit request controller = hf.submit( arguments={ "prompt": "A futuristic city at sunset, flying cars", "model": "sora_2", "duration": 5 } ) # Poll for status for status in hf.poll_request_status(controller.request_id): if isinstance(status, hf.Queued): print(f"Queue position: {status.position}") elif isinstance(status, hf.InProgress): print(f"Progress: {status.progress}%") elif isinstance(status, hf.Completed): print(f"Done! URL: {status.output_url}") break elif isinstance(status, hf.Failed): print(f"Error: {status.error}") break
Webhook Notifications:
result = hf.submit( arguments={...}, webhook_url="https://your-server.com/webhook" )
SDK Methods Reference:
| Method | Description |
|---|---|
| Submit and wait for completion |
| Submit and get controller for tracking |
| Check request status |
| Get completed result |
| Cancel queued request |
| Upload raw bytes |
| Upload from file path |
| Upload PIL Image |
All methods have
_async variants for async/await usage.
Status Types:
- Waiting in queueQueued
- Currently generatingInProgress
- Done, output availableCompleted
- Generation failedFailed
- Content flaggedNSFW
- User cancelledCancelled
Camera Presets Complete Reference
Basic Movements
| Preset | Motion | Use Case |
|---|---|---|
| Static | No movement | Dialogue, portraits |
| Dolly In | Camera moves toward subject | Emphasis, intimacy |
| Dolly Out | Camera moves away | Reveal environment |
| Dolly Zoom | Zoom opposite of dolly | Vertigo effect |
| Pan Left/Right | Horizontal rotation | Scene scan |
| Tilt Up/Down | Vertical rotation | Reveal height |
Dynamic Movements
| Preset | Motion | Use Case |
|---|---|---|
| Whip Pan | Fast horizontal snap | Scene transition |
| Crash Zoom | Rapid zoom in | Shock, emphasis |
| Push In | Slow move toward | Building tension |
| Pull Out | Slow move away | Context reveal |
| Arc Shot | Semi-circular movement | Hero shots |
| 360 Rotation | Full circle around subject | Product showcase |
Aerial/Specialty
| Preset | Motion | Use Case |
|---|---|---|
| FPV Drone | First-person flying | Action, extreme sports |
| Crane Shot | Vertical lift | Establishing shots |
| Helicopter | Aerial sweep | Landscapes |
| Bullet Time | Frozen time orbit | Action freeze |
| Steadicam | Smooth follow | Walking/talking |
| Handheld | Slight shake | Documentary feel |
Experimental (SOUL)
| Preset | Description |
|---|---|
| Escalator | Moving escalator POV |
| Library | Bookshelf tracking |
| Gallery | Art museum walk |
| Street View | Urban street level |
| Subway | Metro station |
| Mt. Fuji | Mountain vista |
| Sunset Beach | Beach sunset |
| Flight Mode | Airplane window |
| Angel Wings | Ethereal floating |
| CCTV | Security camera |
Pricing & Credits
| Plan | Monthly Cost | Credits | Key Features |
|---|---|---|---|
| Free | $0 | Daily limit | 720p, watermark, basic presets |
| Basic | $9/mo | 150/mo | 1080p, no watermark |
| Pro | $29/mo | 600/mo | Faster queue, more models |
| Ultimate | $49/mo | 1,200/mo | Unlimited Nano Banana Pro |
| Creator | $249/mo | 6,000/mo | Priority rendering |
| Studio | Custom | Custom | Team collaboration, API |
Credit Packs: One-time purchases valid for 90 days
Model Credit Costs (approximate):
| Model | Credits/Generation |
|---|---|
| Nano Banana | 5-10 |
| WAN 2.5 | 15-25 |
| Kling 2.6 | 20-30 |
| Sora 2 | 30-50 |
| Sora 2 Pro | 50-100 |
| Veo 3.1 | 25-40 |
Best Practices
Image Quality
- Use 1024x1024 or higher resolution
- Avoid blurry or compressed images
- Good lighting in source = better output
- Centered subjects work best for most presets
Prompt Engineering (Text-to-Video)
Good: "A golden retriever runs through autumn leaves in a forest, morning sunlight, slow motion" Better: "A golden retriever runs joyfully through scattered autumn leaves in a sun-dappled forest, warm morning light filtering through trees, slow motion, cinematic depth of field, camera tracking shot"
Soul ID Optimization
- Minimum 10 reference images
- Include full face, 3/4 angle, profile
- Vary expressions (smile, neutral, talking)
- Consistent lighting across images
- Avoid sunglasses, hats (unless character feature)
VFX Integration
- Match effect intensity to content tone
- Preview before generating full video
- Layer camera movement AFTER effect selection
- Some effects work better on certain subject types
API Efficiency
- Batch similar requests together
- Use webhooks for long generations
- Cache frequently-used image uploads
- Handle status polling with backoff
Troubleshooting
| Issue | Cause | Solution |
|---|---|---|
| Character looks different | Soul ID needs more references | Add 5+ more varied photos |
| Video is too short | Model/credit limitation | Upgrade plan or use longer model |
| VFX looks artificial | Low source quality | Use higher res source image |
| Generation failed | NSFW detection or model error | Rephrase prompt, change source |
| API timeout | Server load | Retry with exponential backoff |
| Queue position not moving | High demand | Check Higgsfield status page |
Model Selection Guide
| Need | Recommended Model | Why |
|---|---|---|
| Fastest generation | Nano Banana | 5-10 seconds |
| Best quality | Sora 2 Pro | Highest fidelity |
| Talking heads | Veo 3.1 + UGC Builder | Best lip-sync |
| Audio sync | WAN 2.5 | Native audio support |
| Long form (8+ sec) | Sora 2 | Extended generation |
| Specific style | Kling 2.6 | Style consistency |
| Budget conscious | WAN 2.5 | Good quality/cost ratio |
Integration Examples
Node.js Wrapper (fetch-based)
async function generateVideo(imageUrl, preset) { const response = await fetch('https://cloud.higgsfield.ai/api/generate', { method: 'POST', headers: { 'Authorization': `Bearer ${process.env.HF_KEY}`, 'Content-Type': 'application/json' }, body: JSON.stringify({ image: imageUrl, camera_preset: preset, model: 'wan_2.5' }) }); return response.json(); }
Next.js API Route
// app/api/higgsfield/route.ts import { NextRequest, NextResponse } from 'next/server' export async function POST(request: NextRequest) { const { imageUrl, preset } = await request.json() // Forward to Higgsfield API const result = await generateVideo(imageUrl, preset) return NextResponse.json(result) }
Webhook Handler
// app/api/higgsfield-webhook/route.ts export async function POST(request: NextRequest) { const event = await request.json() if (event.status === 'completed') { // Save video URL to database await db.videos.update({ where: { requestId: event.request_id }, data: { outputUrl: event.output_url, status: 'ready' } }) } return NextResponse.json({ received: true }) }
Resources
- Higgsfield Platform
- Higgsfield Cloud API
- Python SDK (GitHub)
- Camera Controls Guide
- VFX Effects Library
- Soul ID Documentation
- WAN 2.5 Features
- Sora 2 Integration
This skill is maintained by ID8Labs. Last updated: 2026-01-19