Skip to content

Add GLM 4.6 and LongCat Flash Thinking for Chutes AI provider #8425

@leakless21

Description

@leakless21

Problem (one or two sentences)

The Roo Code extension currently lacks support for the zai-org/GLM-4.6-FP8 (200k context window) and meituan-longcat/LongCat-Flash-Thinking-FP8 (128k context window) models on the Chutes AI provider. These model performs on par or very close to with SOTA models while being much faster and cheaper.

Context (who is affected and when)

This affects all Roo Code users who needs efficient models for complex reasoning, coding, and creative tasks. There are no options for these models in Roo Code/ Chutes AI.

Desired behavior (conceptual, not technical)

The user should see and be able to select zai-org/GLM-4.6-FP8 and meituan-longcat/LongCat-Flash-Thinking-FP8 models inside Roo Code, specifically Chutes AI provider

Constraints / preferences (optional)

No response

Request checklist

  • I've searched existing Issues and Discussions for duplicates
  • This describes a specific problem with clear context and impact

Roo Code Task Links (optional)

No response

Acceptance criteria (optional)

No response

Proposed approach (optional)

No response

Trade-offs / risks (optional)

No response

Metadata

Metadata

Assignees

No one assigned

    Labels

    EnhancementNew feature or requestIssue - In ProgressSomeone is actively working on this. Should link to a PR soon.feature requestFeature request, not a bug

    Type

    No type

    Projects

    Status

    Done

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions