Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tool Calling for firefunction-v2 GGUF #316

Open
odrobnik opened this issue Jan 16, 2025 · 1 comment
Open

Tool Calling for firefunction-v2 GGUF #316

odrobnik opened this issue Jan 16, 2025 · 1 comment
Assignees

Comments

@odrobnik
Copy link

Which version of LM Studio?
LM Studio 0.3.6

Which operating system?
macOS

What is the bug?
Sending a chat completion request to LM Studio for firefunction-v2, including Tools results in

Error: received prediction-error t.fromSerializedError (/Applications/LM Studio.app/Contents/Resources/app/.webpack/main/index.js:382:62222) at m.o (/Applications/LM Studio.app/Contents/Resources/app/.webpack/main/index.js:25:155460) at m.emit (node:events:519:28) at m.onChildMessage (/Applications/LM Studio.app/Contents/Resources/app/.webpack/main/index.js:25:140859) at m.onChildMessage (/Applications/LM Studio.app/Contents/Resources/app/.webpack/main/index.js:25:164118) at ForkUtilityProcess.<anonymous> (/Applications/LM Studio.app/Contents/Resources/app/.webpack/main/index.js:25:140194) at ForkUtilityProcess.emit (node:events:519:28) at ForkUtilityProcess.a.emit (node:electron/js2c/browser_init:2:71823) - Caused By: SyntaxError: Unknown statement type: Identifier at /Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:33:115985 at i (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:33:116050) at R (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:33:120440) at new se (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:33:137373) at /Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:10:36458 at async t.formatPrompt (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:10:34152) at async t.getContextOverflowPolicy (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:10:49555) at async t.temp_createLlamaPredictionArgs (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:8:366412) at async t.LLMEngineWrapper.predictTokens (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:10:26226) at async Object.predictTokens (/Applications/LM Studio.app/Contents/Resources/app/.webpack/lib/llmworker.js:15:7181)

Screenshots

Image

Additional Info

For reference: tool calling works ok using the same model and ollama.

firefunction-v2 is a fine tune of Llama 3 to support multiple tool calls in parallel by making one JSON with multiple function calls.

@yagil
Copy link
Member

yagil commented Jan 16, 2025

Thanks @odrobnik ! We're investigating

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants