skills/imfa-solutions/skills/convex-rate-limiter

convex-rate-limiter

SKILL.md

Convex Rate Limiter

@convex-dev/rate-limiter — Type-safe, transactional, application-level rate limiting for Convex.

Installation

npm install @convex-dev/rate-limiter

Register the component in convex/convex.config.ts:

import { defineApp } from "convex/server";
import rateLimiter from "@convex-dev/rate-limiter/convex.config.js";

const app = defineApp();
app.use(rateLimiter);
export default app;

Setup — Define Named Rate Limits

import { RateLimiter, MINUTE, HOUR, SECOND } from "@convex-dev/rate-limiter";
import { components } from "./_generated/api";

const rateLimiter = new RateLimiter(components.rateLimiter, {
  // Fixed window — hard quota that resets each period
  freeTrialSignUp: { kind: "fixed window", rate: 100, period: HOUR },

  // Token bucket — smooth traffic with burst allowance
  sendMessage: { kind: "token bucket", rate: 10, period: MINUTE, capacity: 3 },

  // Failed login throttle
  failedLogins: { kind: "token bucket", rate: 10, period: HOUR },
});
  • period is in milliseconds (SECOND = 1000, MINUTE = 60000, HOUR = 3600000).
  • Multiple RateLimiter instances allowed; config keys must not overlap.

Strategy Selection

Strategy Behavior Best for
Token bucket Tokens refill continuously; unused tokens accumulate up to capacity User actions, API calls, LLM tokens — smooth traffic, allow bursts
Fixed window All tokens granted at period start; hard reset each period Daily/hourly quotas, signup caps, hard API limits

Token bucket config:

{ kind: "token bucket", rate: 10, period: MINUTE, capacity: 3 }
// capacity = max burst tokens (optional, defaults to rate)

Fixed window config:

{ kind: "fixed window", rate: 100, period: HOUR }
// start: optional custom epoch; random by default to prevent thundering herd

Core API

limit() — Consume tokens

const { ok, retryAfter } = await rateLimiter.limit(ctx, "sendMessage", {
  key: userId,       // Per-entity key (omit for global limit)
  count: 1,          // Tokens to consume (default 1)
  throws: false,     // Set true to auto-throw ConvexError
});
if (!ok) throw new Error(`Rate limited. Retry in ${Math.ceil(retryAfter / 1000)}s`);

check() — Query without consuming (safe in queries)

const { ok, retryAfter } = await rateLimiter.check(ctx, "sendMessage", {
  key: userId,
});

reset() — Clear a rate limit

await rateLimiter.reset(ctx, "failedLogins", { key: email });

throws: true — Auto-throw pattern

import { isRateLimitError } from "@convex-dev/rate-limiter";

await rateLimiter.limit(ctx, "sendMessage", { key: userId, throws: true });
// Throws ConvexError with data: { kind, name, retryAfter }

Catch with isRateLimitError(error) to inspect .data.retryAfter.

Usage Patterns

Global rate limit (no key)

export const signUp = mutation({
  args: { email: v.string() },
  handler: async (ctx, args) => {
    await rateLimiter.limit(ctx, "freeTrialSignUp", { throws: true });
    await ctx.db.insert("users", { email: args.email });
  },
});

Per-user rate limit

export const sendMessage = mutation({
  args: { text: v.string() },
  handler: async (ctx, args) => {
    const user = await ctx.auth.getUserIdentity();
    await rateLimiter.limit(ctx, "sendMessage", {
      key: user?.subject,
      throws: true,
    });
    await ctx.db.insert("messages", { userId: user?.subject, text: args.text });
  },
});

Failed login throttle with reset on success

export const login = mutation({
  args: { email: v.string(), password: v.string() },
  handler: async (ctx, args) => {
    await rateLimiter.limit(ctx, "failedLogins", {
      key: args.email,
      throws: true,
    });
    const success = await verifyCredentials(ctx, args);
    if (success) {
      await rateLimiter.reset(ctx, "failedLogins", { key: args.email });
    }
    return success;
  },
});

Consume multiple tokens at once

await rateLimiter.limit(ctx, "llmTokens", {
  key: userId,
  count: estimateTokens(prompt),
  throws: true,
});

Critical Rules

  1. Always pass key for per-entity limits. Omitting key makes it a global singleton.
  2. Always surface retryAfter to the client — don't just say "rate limited".
  3. Rate limit changes are transactional — they roll back if the mutation fails.
  4. Use throws: true for cleaner code; catch with isRateLimitError().
  5. Reset wisely — reset on success (e.g., login), never on every request.

Common Pitfalls

// BAD: Global limit when you want per-user
await rateLimiter.limit(ctx, "sendMessage");

// GOOD: Per-user limit
await rateLimiter.limit(ctx, "sendMessage", { key: userId });
// BAD: Ignoring retryAfter
const { ok } = await rateLimiter.limit(ctx, "action");
if (!ok) throw new Error("Rate limited");

// GOOD: Tell user when to retry
const { ok, retryAfter } = await rateLimiter.limit(ctx, "action");
if (!ok) throw new Error(`Wait ${Math.ceil(retryAfter / 1000)} seconds`);

Best Practices Summary

Practice Guidance
Key design User ID for per-user, IP for anonymous, team ID for team-wide, composites like ${teamId}:${userId}
Capacity Token bucket: capacity = rate * burst_multiplier. Fixed window: defaults to rate
Start simple Add sharding/reservation only when needed
Error handling Use throws: true + isRateLimitError()
Thundering herd Fixed window uses random start by default; add jitter to retry times

Advanced Topics

Common Patterns

Weekly Installs
14
GitHub Stars
1
First Seen
Feb 24, 2026
Installed on
opencode14
gemini-cli14
antigravity14
claude-code14
github-copilot14
codex14