I find it very interesting that the LLM is told so little details but seems to just intuitively understand based on the english words used for the tool name and function arguments.
I know from earlier discussions that this is partially because many LLMs have been fine tuned on function calling, however the model providers don’t share this training dataset unfortunately. I think models that haven’t been fine tuned can still do function calling with careful instructions in their system prompt but are much worse at it.
Thank you for comments that help with learning and understanding MCP and tools better.