Fine-tuned on a mix of reasoning data and FunctionGemma-related function-calling data, then converted for llama.cpp/GGUF runtimes.

It’s still a Qwen-family model, but the tuning pushes it more toward structured responses, tool-use style behavior, and action-oriented prompting.