mirror of
https://gitea.toothfairyai.com/ToothFairyAI/tf_code.git
synced 2026-04-15 05:04:50 +00:00
feat: add litellmProxy provider option for explicit LiteLLM compatibility (#8658)
Co-authored-by: Mark Henderson <Mark.Henderson99@hotmail.com> Co-authored-by: Aiden Cline <63023139+rekram1-node@users.noreply.github.com>
This commit is contained in:
90
packages/opencode/test/session/llm.test.ts
Normal file
90
packages/opencode/test/session/llm.test.ts
Normal file
@@ -0,0 +1,90 @@
|
||||
import { describe, expect, test } from "bun:test"
|
||||
import { LLM } from "../../src/session/llm"
|
||||
import type { ModelMessage } from "ai"
|
||||
|
||||
describe("session.llm.hasToolCalls", () => {
|
||||
test("returns false for empty messages array", () => {
|
||||
expect(LLM.hasToolCalls([])).toBe(false)
|
||||
})
|
||||
|
||||
test("returns false for messages with only text content", () => {
|
||||
const messages: ModelMessage[] = [
|
||||
{
|
||||
role: "user",
|
||||
content: [{ type: "text", text: "Hello" }],
|
||||
},
|
||||
{
|
||||
role: "assistant",
|
||||
content: [{ type: "text", text: "Hi there" }],
|
||||
},
|
||||
]
|
||||
expect(LLM.hasToolCalls(messages)).toBe(false)
|
||||
})
|
||||
|
||||
test("returns true when messages contain tool-call", () => {
|
||||
const messages = [
|
||||
{
|
||||
role: "user",
|
||||
content: [{ type: "text", text: "Run a command" }],
|
||||
},
|
||||
{
|
||||
role: "assistant",
|
||||
content: [
|
||||
{
|
||||
type: "tool-call",
|
||||
toolCallId: "call-123",
|
||||
toolName: "bash",
|
||||
},
|
||||
],
|
||||
},
|
||||
] as ModelMessage[]
|
||||
expect(LLM.hasToolCalls(messages)).toBe(true)
|
||||
})
|
||||
|
||||
test("returns true when messages contain tool-result", () => {
|
||||
const messages = [
|
||||
{
|
||||
role: "tool",
|
||||
content: [
|
||||
{
|
||||
type: "tool-result",
|
||||
toolCallId: "call-123",
|
||||
toolName: "bash",
|
||||
},
|
||||
],
|
||||
},
|
||||
] as ModelMessage[]
|
||||
expect(LLM.hasToolCalls(messages)).toBe(true)
|
||||
})
|
||||
|
||||
test("returns false for messages with string content", () => {
|
||||
const messages: ModelMessage[] = [
|
||||
{
|
||||
role: "user",
|
||||
content: "Hello world",
|
||||
},
|
||||
{
|
||||
role: "assistant",
|
||||
content: "Hi there",
|
||||
},
|
||||
]
|
||||
expect(LLM.hasToolCalls(messages)).toBe(false)
|
||||
})
|
||||
|
||||
test("returns true when tool-call is mixed with text content", () => {
|
||||
const messages = [
|
||||
{
|
||||
role: "assistant",
|
||||
content: [
|
||||
{ type: "text", text: "Let me run that command" },
|
||||
{
|
||||
type: "tool-call",
|
||||
toolCallId: "call-456",
|
||||
toolName: "read",
|
||||
},
|
||||
],
|
||||
},
|
||||
] as ModelMessage[]
|
||||
expect(LLM.hasToolCalls(messages)).toBe(true)
|
||||
})
|
||||
})
|
||||
Reference in New Issue
Block a user