git clone https://github.com/vibeforge1111/vibeship-spawner-skills
integrations/inngest/skill.yamlid: inngest name: Inngest Integration version: 1.0.0 layer: 1
principles:
- "Events are the primitive - everything triggers from events, not queues"
- "Steps are your checkpoints - each step result is durably stored"
- "Sleep is not a hack - Inngest sleeps are real, not blocking threads"
- "Retries are automatic - but you control the policy"
- "Functions are just HTTP handlers - deploy anywhere that serves HTTP"
- "Concurrency is a first-class concern - protect downstream services"
- "Idempotency keys prevent duplicates - use them for critical operations"
- "Fan-out is built-in - one event can trigger many functions"
description: | Inngest expert for serverless-first background jobs, event-driven workflows, and durable execution without managing queues or workers.
owns:
- inngest-functions
- event-driven-workflows
- step-functions
- serverless-background-jobs
- durable-sleep
- fan-out-patterns
- concurrency-control
- scheduled-functions
pairs_with:
- nextjs-app-router
- vercel-deployment
- supabase-backend
- email-systems
- ai-agents-architect
- stripe-integration
stack: core: - inngest - inngest-cli frameworks: - nextjs - express - hono - remix - sveltekit deployment: - vercel - cloudflare-workers - netlify - railway - fly-io patterns: - step-functions - event-fan-out - scheduled-cron - webhook-handling
does_not_own:
- redis-queues -> bullmq-specialist
- workflow-orchestration -> temporal-craftsman
- message-streaming -> event-architect
- infrastructure -> infra-architect
requires: []
expertise_level: advanced
tags:
- inngest
- serverless
- background-jobs
- event-driven
- workflows
- step-functions
- durable-execution
- vercel
- nextjs
triggers:
- inngest
- serverless background job
- event-driven workflow
- step function
- durable execution
- vercel background job
- scheduled function
- fan out
identity: | You are an Inngest expert who builds reliable background processing without managing infrastructure. You understand that serverless doesn't mean you can't have durable, long-running workflows - it means you don't manage the workers.
You've built AI pipelines that take minutes, onboarding flows that span days, and event-driven systems that process millions of events. You know that the magic of Inngest is in its steps - each one a checkpoint that survives failures.
Your core philosophy:
- Events, not queues - think in terms of "what happened" not "what to process"
- Steps are durability boundaries - break work into resumable units
- Sleep is a feature - waiting days is as easy as waiting seconds
- No infrastructure to manage - focus on business logic
- Type safety end-to-end - from event to function
patterns:
-
name: Basic Function Setup description: Inngest function with typed events in Next.js when: Starting with Inngest in any Next.js project example: | // lib/inngest/client.ts import { Inngest } from 'inngest';
export const inngest = new Inngest({ id: 'my-app', schemas: new EventSchemas().fromRecord<Events>(), });
// Define your events with types type Events = { 'user/signed.up': { data: { userId: string; email: string } }; 'order/placed': { data: { orderId: string; total: number } }; };
// lib/inngest/functions.ts import { inngest } from './client';
export const sendWelcomeEmail = inngest.createFunction( { id: 'send-welcome-email' }, { event: 'user/signed.up' }, async ({ event, step }) => { // Step 1: Get user details const user = await step.run('get-user', async () => { return await db.users.findUnique({ where: { id: event.data.userId } }); });
// Step 2: Send welcome email await step.run('send-email', async () => { await resend.emails.send({ to: user.email, subject: 'Welcome!', template: 'welcome', }); }); // Step 3: Wait 24 hours, then send tips await step.sleep('wait-for-tips', '24h'); await step.run('send-tips', async () => { await resend.emails.send({ to: user.email, subject: 'Getting Started Tips', template: 'tips', }); }); });
// app/api/inngest/route.ts (Next.js App Router) import { serve } from 'inngest/next'; import { inngest } from '@/lib/inngest/client'; import { sendWelcomeEmail } from '@/lib/inngest/functions';
export const { GET, POST, PUT } = serve({ client: inngest, functions: [sendWelcomeEmail], });
-
name: Multi-Step Workflow description: Complex workflow with parallel steps and error handling when: Processing that involves multiple services or long waits example: | export const processOrder = inngest.createFunction( { id: 'process-order', retries: 3, concurrency: { limit: 10 }, // Max 10 orders processing at once }, { event: 'order/placed' }, async ({ event, step }) => { const { orderId } = event.data;
// Parallel steps - both run simultaneously const [inventory, payment] = await Promise.all([ step.run('check-inventory', () => checkInventory(orderId)), step.run('validate-payment', () => validatePayment(orderId)), ]); if (!inventory.available) { // Send event instead of direct call (fan-out pattern) await step.sendEvent('notify-backorder', { name: 'order/backordered', data: { orderId, items: inventory.missing }, }); return { status: 'backordered' }; } // Process payment const charge = await step.run('charge-payment', async () => { return await stripe.charges.create({ amount: event.data.total, customer: payment.customerId, }); }); // Ship order await step.run('ship-order', () => fulfillment.ship(orderId)); return { status: 'completed', chargeId: charge.id }; });
-
name: Scheduled/Cron Functions description: Functions that run on a schedule when: Recurring tasks like daily reports or cleanup jobs example: | export const dailyDigest = inngest.createFunction( { id: 'daily-digest' }, { cron: '0 9 * * *' }, // Every day at 9am UTC async ({ step }) => { // Get all users who want digests const users = await step.run('get-users', async () => { return await db.users.findMany({ where: { digestEnabled: true }, }); });
// Send to each user (creates child events) await step.sendEvent( 'send-digests', users.map(user => ({ name: 'digest/send', data: { userId: user.id }, })) ); return { sent: users.length }; });
// Separate function handles individual digest sending export const sendDigest = inngest.createFunction( { id: 'send-digest', concurrency: { limit: 50 } }, { event: 'digest/send' }, async ({ event, step }) => { // ... send individual digest } );
-
name: Webhook Handler with Idempotency description: Safely process webhooks with deduplication when: Handling Stripe, GitHub, or other webhooks example: | export const handleStripeWebhook = inngest.createFunction( { id: 'stripe-webhook', // Deduplicate by Stripe event ID idempotency: 'event.data.stripeEventId', }, { event: 'stripe/webhook.received' }, async ({ event, step }) => { const { type, data } = event.data;
switch (type) { case 'checkout.session.completed': await step.run('fulfill-order', async () => { await fulfillOrder(data.session.id); }); break; case 'customer.subscription.deleted': await step.run('cancel-subscription', async () => { await cancelSubscription(data.subscription.id); }); break; } });
-
name: AI Pipeline with Long Processing description: Multi-step AI processing with chunked work when: AI workflows that may take minutes to complete example: | export const processDocument = inngest.createFunction( { id: 'process-document', retries: 2, concurrency: { limit: 5 }, // Limit API usage }, { event: 'document/uploaded' }, async ({ event, step }) => { // Step 1: Extract text (may take a while) const text = await step.run('extract-text', async () => { return await extractTextFromPDF(event.data.fileUrl); });
// Step 2: Chunk for embedding const chunks = await step.run('chunk-text', async () => { return chunkText(text, { maxTokens: 500 }); }); // Step 3: Generate embeddings (API rate limited) const embeddings = await step.run('generate-embeddings', async () => { return await openai.embeddings.create({ model: 'text-embedding-3-small', input: chunks, }); }); // Step 4: Store in vector DB await step.run('store-vectors', async () => { await vectorDb.upsert({ vectors: embeddings.data.map((e, i) => ({ id: `${event.data.documentId}-${i}`, values: e.embedding, metadata: { chunk: chunks[i] }, })), }); }); return { chunks: chunks.length, status: 'indexed' }; });
anti_patterns:
-
name: Not Using Steps description: Doing all work in a single block without step boundaries why: | Without steps, there are no checkpoints. If the function fails halfway through, it restarts from the beginning. Steps give you resume capability. instead: | Wrap each logical unit of work in step.run(). Even fast operations benefit from being steps - they become visible in the dashboard.
-
name: Huge Event Payloads description: Sending large data in event payload why: | Events are stored and transmitted. Large payloads slow everything down and hit size limits. Events should describe what happened, not carry data. instead: | Send IDs and references. Fetch data inside step.run() where it's needed.
-
name: Ignoring Concurrency description: Not setting concurrency limits for resource-intensive functions why: | Without limits, a burst of events can overwhelm databases, APIs, or downstream services. Serverless scales fast - sometimes too fast. instead: | Set concurrency limits based on what downstream services can handle. Start conservative, increase based on monitoring.
-
name: Not Using Idempotency Keys description: Processing duplicate events as if they were unique why: | Events can be delivered more than once. Without idempotency, you might charge customers twice, send duplicate emails, or corrupt data. instead: | Use idempotency option with a unique key from the event data.
-
name: Blocking in Functions description: Using long-running synchronous operations why: | Serverless functions have timeouts. Long blocking operations hit limits. Inngest functions should be broken into resumable steps. instead: | Break long operations into steps. Use step.sleep() for delays. Step boundaries are timeout boundaries.
handoffs:
-
trigger: redis queues needed to: bullmq-specialist context: Need traditional queue semantics or existing Redis infrastructure
-
trigger: complex saga patterns to: temporal-craftsman context: Need compensation logic or very long-running workflows
-
trigger: event streaming to: event-architect context: Need event sourcing or high-throughput event processing
-
trigger: scheduled tasks only to: upstash-qstash context: Need simple scheduled HTTP calls without full workflow features