I think without fine tuning it on returning function calls you wonβt get any good results. Maybe trying it with validating the input and result again in a loop could do the job, but otherwise we need to wait for fine tuned models.
LocalLLaMA
Community to discuss about Llama, the family of large language models created by Meta AI.
Yes good call π I'm gonna try that now.
I'm gonna add a system message to the chat history part of the prompt saying that the command is invalid and see if it corrects itself in the next iteration of the loop.
This could add a bandage π©Ή on the issue for now. Allowing it to seemlessly loop until a task is complete. Until I can find a better prompt or model.
So far, dolphin-mistral 2.1 7b is what I'm using ATM.
Use langchain with tools
Not a bad idea but I'm not coding in Python.
Because I hate myself, I'm writing in C# π«
Also, I want to use a few libraries as possible.
As a last resort I may use langchain. Or just look at the source and see how they force a model into function calling if it's possible.