Skip to content

feat: add ratelimit and validation to llm search#3311

Merged
mcstepp merged 2 commits intomainfrom
add-validation-to-llm-saerch
Jun 11, 2025
Merged

feat: add ratelimit and validation to llm search#3311
mcstepp merged 2 commits intomainfrom
add-validation-to-llm-saerch

Conversation

@ogzhanolguncu
Copy link
Contributor

@ogzhanolguncu ogzhanolguncu commented Jun 10, 2025

What does this PR do?

This PR adds stricter ratelimit for LLM and also adds input length checks to prevent abuse.
Fixes # (issue)

If there is not an issue for this, please create one first. This is used to tracking purposes and also helps use understand why this PR exists

Type of change

  • Bug fix (non-breaking change which fixes an issue)
  • Chore (refactoring code, technical debt, workflow improvements)
  • Enhancement (small improvements)
  • New feature (non-breaking change which adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • This change requires a documentation update

How should this be tested?

  • Test A
  • Test B

Checklist

Required

  • Filled out the "How to test" section in this PR
  • Read Contributing Guide
  • Self-reviewed my own code
  • Commented on my code in hard-to-understand areas
  • Ran pnpm build
  • Ran pnpm fmt
  • Checked for warnings, there are none
  • Removed all console.logs
  • Merged the latest changes from main onto my branch with git pull origin main
  • My changes don't cause any responsiveness issues

Appreciated

  • If a UI change was made: Added a screen recording or screenshots to this PR
  • Updated the Unkey Docs if changes were necessary

Summary by CodeRabbit

  • New Features

    • Added a character limit of 120 for the LLM search input field.
    • Introduced rate limiting and input validation for LLM-related queries to enhance stability and prevent excessive usage.
  • Improvements

    • Adjusted LLM query handling to enforce stricter input validation and improved error messaging.
    • Refined LLM search parameters for more consistent results.

@changeset-bot
Copy link

changeset-bot bot commented Jun 10, 2025

⚠️ No Changeset found

Latest commit: 6f9c0a3

Merging this PR will not cause a version bump for any packages. If these changes should not result in a new version, you're good to go. If these changes should result in a version bump, you need to add a changeset.

This PR includes no changesets

When changesets are added to this PR, you'll see the packages that this PR includes changesets for and the associated semver types

Click here to learn what changesets are, and how to add one.

Click here if you're a maintainer who wants to add a changeset to this PR

@vercel
Copy link

vercel bot commented Jun 10, 2025

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Name Status Preview Comments Updated (UTC)
dashboard ✅ Ready (Inspect) Visit Preview 💬 Add feedback Jun 10, 2025 2:54pm
1 Skipped Deployment
Name Status Preview Comments Updated (UTC)
engineering ⬜️ Ignored (Inspect) Visit Preview Jun 10, 2025 2:54pm

@coderabbitai
Copy link
Contributor

coderabbitai bot commented Jun 10, 2025

Warning

Rate limit exceeded

@ogzhanolguncu has exceeded the limit for the number of commits or files that can be reviewed per hour. Please wait 12 minutes and 44 seconds before requesting another review.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

📥 Commits

Reviewing files that changed from the base of the PR and between 350ed7b and 6f9c0a3.

📒 Files selected for processing (1)
  • apps/dashboard/lib/trpc/routers/logs/llm-search/index.ts (2 hunks)
📝 Walkthrough

Walkthrough

A new middleware, withLlmAccess, was introduced to enforce rate limiting and query validation for LLM search requests. The LLM search procedure was updated to use this middleware and now expects the query string from context instead of input. The search input field is limited to 120 characters, and LLM request parameters were slightly adjusted.

Changes

File(s) Change Summary
apps/dashboard/components/logs/llm-search/components/search-input.tsx Added LLM_LIMITS_MAX_QUERY_LENGTH constant and set input field's maxLength to 120.
apps/dashboard/lib/trpc/routers/logs/llm-search/index.ts Switched middleware from rate limiting to withLlmAccess; input schema changed; query from context.
apps/dashboard/lib/trpc/routers/logs/llm-search/utils.ts Reduced OpenAI temperature to 0.1; removed top_p, frequency_penalty, presence_penalty.
apps/dashboard/lib/trpc/trpc.ts Added LLM_LIMITS constant and new withLlmAccess middleware for LLM query rate limiting/validation.

Sequence Diagram(s)

sequenceDiagram
    participant User
    participant SearchInput
    participant TRPC_Server
    participant withLlmAccess
    participant LLM_Procedure
    participant OpenAI

    User->>SearchInput: Enter search query (max 120 chars)
    SearchInput->>TRPC_Server: Submit query
    TRPC_Server->>withLlmAccess: Apply rate limiting & validate query
    withLlmAccess-->>TRPC_Server: Pass validated query in context
    TRPC_Server->>LLM_Procedure: Call procedure with validated query
    LLM_Procedure->>OpenAI: Send request (temperature 0.1)
    OpenAI-->>LLM_Procedure: Return results
    LLM_Procedure-->>TRPC_Server: Return results
    TRPC_Server-->>SearchInput: Return results
    SearchInput-->>User: Display search results
Loading
✨ Finishing Touches
  • 📝 Generate Docstrings

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share
🪧 Tips

Chat

There are 3 ways to chat with CodeRabbit:

  • Review comments: Directly reply to a review comment made by CodeRabbit. Example:
    • I pushed a fix in commit <commit_id>, please review it.
    • Explain this complex logic.
    • Open a follow-up GitHub issue for this discussion.
  • Files and specific lines of code (under the "Files changed" tab): Tag @coderabbitai in a new review comment at the desired location with your query. Examples:
    • @coderabbitai explain this code block.
    • @coderabbitai modularize this function.
  • PR comments: Tag @coderabbitai in a new PR comment to ask questions about the PR branch. For the best results, please provide a very specific query, as very limited context is provided in this mode. Examples:
    • @coderabbitai gather interesting stats about this repository and render them as a table. Additionally, render a pie chart showing the language distribution in the codebase.
    • @coderabbitai read src/utils.ts and explain its main purpose.
    • @coderabbitai read the files in the src/scheduler package and generate a class diagram using mermaid and a README in the markdown format.
    • @coderabbitai help me debug CodeRabbit configuration file.

Support

Need help? Create a ticket on our support page for assistance with any issues or questions.

Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments.

CodeRabbit Commands (Invoked using PR comments)

  • @coderabbitai pause to pause the reviews on a PR.
  • @coderabbitai resume to resume the paused reviews.
  • @coderabbitai review to trigger an incremental review. This is useful when automatic reviews are disabled for the repository.
  • @coderabbitai full review to do a full review from scratch and review all the files again.
  • @coderabbitai summary to regenerate the summary of the PR.
  • @coderabbitai generate docstrings to generate docstrings for this PR.
  • @coderabbitai generate sequence diagram to generate a sequence diagram of the changes in this PR.
  • @coderabbitai resolve resolve all the CodeRabbit review comments.
  • @coderabbitai configuration to show the current CodeRabbit configuration for the repository.
  • @coderabbitai help to get help.

Other keywords and placeholders

  • Add @coderabbitai ignore anywhere in the PR description to prevent this PR from being reviewed.
  • Add @coderabbitai summary to generate the high-level summary at a specific location in the PR description.
  • Add @coderabbitai anywhere in the PR title to generate the title automatically.

CodeRabbit Configuration File (.coderabbit.yaml)

  • You can programmatically configure CodeRabbit by adding a .coderabbit.yaml file to the root of your repository.
  • Please see the configuration documentation for more information.
  • If your editor has YAML language server enabled, you can add the path at the top of this file to enable auto-completion and validation: # yaml-language-server: $schema=https://coderabbit.ai/integrations/schema.v2.json

Documentation and Community

  • Visit our Documentation for detailed information on how to use CodeRabbit.
  • Join our Discord Community to get help, request features, and share feedback.
  • Follow us on X/Twitter for updates and announcements.

@github-actions
Copy link
Contributor

github-actions bot commented Jun 10, 2025

Thank you for following the naming conventions for pull request titles! 🙏

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

📜 Review details

Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between 17a7355 and 350ed7b.

📒 Files selected for processing (4)
  • apps/dashboard/components/logs/llm-search/components/search-input.tsx (2 hunks)
  • apps/dashboard/lib/trpc/routers/logs/llm-search/index.ts (2 hunks)
  • apps/dashboard/lib/trpc/routers/logs/llm-search/utils.ts (1 hunks)
  • apps/dashboard/lib/trpc/trpc.ts (2 hunks)
🧰 Additional context used
🧬 Code Graph Analysis (2)
apps/dashboard/lib/trpc/routers/logs/llm-search/index.ts (2)
apps/dashboard/lib/trpc/trpc.ts (1)
  • withLlmAccess (155-189)
apps/dashboard/lib/trpc/routers/logs/llm-search/utils.ts (1)
  • getStructuredSearchFromLLM (7-75)
apps/dashboard/lib/trpc/trpc.ts (1)
apps/dashboard/lib/env.ts (1)
  • env (3-49)
⏰ Context from checks skipped due to timeout of 90000ms (8)
  • GitHub Check: Test Packages / Test ./packages/rbac
  • GitHub Check: Test Packages / Test ./packages/hono
  • GitHub Check: Test Packages / Test ./packages/cache
  • GitHub Check: Test Packages / Test ./packages/nextjs
  • GitHub Check: Test Packages / Test ./internal/clickhouse
  • GitHub Check: Test Packages / Test ./internal/billing
  • GitHub Check: autofix
  • GitHub Check: Analyze (javascript-typescript)
🔇 Additional comments (11)
apps/dashboard/lib/trpc/trpc.ts (5)

5-5: LGTM: Zod import added for validation schema.

The import is necessary for the new input validation functionality.


130-136: Well-structured LLM limits configuration.

The constants are clearly defined with reasonable values. Using as const ensures type safety and immutability.


138-145: Proper rate limiter instantiation with environment-based configuration.

The conditional creation and dedicated namespace "trpc_llm" provides good separation from other rate limiters.


147-153: Robust input validation schema with helpful error messages.

The schema correctly trims input and enforces length constraints with user-friendly error messages.


155-189: 🛠️ Refactor suggestion

Consider adding type safety for rawInput validation.

The middleware implementation is solid, but there's a potential runtime issue if rawInput is not an object or lacks the expected structure.

Consider adding a type guard or safer parsing approach:

export const withLlmAccess = () =>
  t.middleware(async ({ next, ctx, rawInput }) => {
    if (llmRatelimit) {
      const response = await llmRatelimit.limit(ctx.user!.id);
      if (!response.success) {
        throw new TRPCError({
          code: "TOO_MANY_REQUESTS",
          message: `LLM rate limit exceeded. You can make ${LLM_LIMITS.RATE_LIMIT} requests per minute.`,
        });
      }
    }

    let validatedInput: z.infer<typeof llmQuerySchema>;
    try {
+     // Ensure rawInput is an object before parsing
+     if (!rawInput || typeof rawInput !== 'object') {
+       throw new TRPCError({
+         code: "BAD_REQUEST",
+         message: "Invalid input format",
+       });
+     }
      validatedInput = llmQuerySchema.parse(rawInput);
    } catch (error) {
      if (error instanceof z.ZodError) {
        const firstError = error.errors[0];
        throw new TRPCError({
          code: "BAD_REQUEST",
          message: firstError?.message || "Invalid query format",
        });
      }
      throw new TRPCError({
        code: "BAD_REQUEST",
        message: "Invalid input format",
      });
    }

    return next({
      ctx: {
        validatedQuery: validatedInput.query,
      },
    });
  });

Likely an incorrect or invalid review comment.

apps/dashboard/lib/trpc/routers/logs/llm-search/utils.ts (1)

22-22: Good adjustment for more deterministic search results.

Reducing the temperature to 0.1 will make the LLM responses more focused and consistent, which is appropriate for structured search query generation.

apps/dashboard/components/logs/llm-search/components/search-input.tsx (1)

43-43: Proper enforcement of input length constraint.

The maxLength attribute correctly enforces the character limit on the frontend, providing immediate feedback to users.

apps/dashboard/lib/trpc/routers/logs/llm-search/index.ts (4)

2-2: Clean import update for new middleware approach.

Properly replaces the generic rate limiting with the specialized LLM access middleware.


16-16: Correct middleware integration.

The withLlmAccess middleware properly handles both rate limiting and input validation for LLM queries.


17-17: Simplified input schema reflects middleware-handled validation.

Removing the query field from input makes sense since the middleware now validates and provides it via context.


19-19: Proper usage of validated query from context.

Correctly accesses the validated query from the middleware context instead of the raw input.

Copy link
Collaborator

@mcstepp mcstepp left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

i like it 👍

@mcstepp mcstepp added this pull request to the merge queue Jun 11, 2025
Merged via the queue into main with commit d9eb0d2 Jun 11, 2025
28 of 30 checks passed
@mcstepp mcstepp deleted the add-validation-to-llm-saerch branch June 11, 2025 17:47
@coderabbitai coderabbitai bot mentioned this pull request Jul 3, 2025
18 tasks
@coderabbitai coderabbitai bot mentioned this pull request Aug 21, 2025
18 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants