Callstack Reviewer supports specifying model, token and endpoint for all LLM calls. Model can be specified either at the top level - which overrides models in all modules - or on per module basis.

Example configuration:

pr_review:
  # specify model globally for all modules
  model:
    model: "gpt-4o"
    token: <model token>
    base_url: <endpoint base url>

  # specify model for specific module
  modules:
    bug_hunter:
      model:
        model: "claude-3-5-sonnet-20241022"
        token: <model token>
        base_url: <endpoint base url>

We highly recommend keeping the default model configuration for the best performance.

Callstack Reviewer supports specifying model, token and endpoint for all LLM calls. Model can be specified either at the top level - which overrides models in all modules - or on per module basis.

Example configuration:

pr_review:
  # specify model globally for all modules
  model:
    model: "gpt-4o"
    token: <model token>
    base_url: <endpoint base url>

  # specify model for specific module
  modules:
    bug_hunter:
      model:
        model: "claude-3-5-sonnet-20241022"
        token: <model token>
        base_url: <endpoint base url>

We highly recommend keeping the default model configuration for the best performance.