Chapter 3. Tool call parser type reference
Tool call parsers determine how AI Inference Server interprets and extracts tool calling information from the model output. Each model has specific output formats for tool calls. You must configure the appropriate parser that matches the expected format for the model. The tool call parser is specified using the --tool-call-parser flag when launching AI Inference Server. AI Inference Server supports multiple tool call parsers for different model families and output formats.
| Parser | Description | Example models |
|---|---|---|
|
| For Nous-Hermes models that use the Hermes tool calling format. |
|
|
| For Mistral models using Mistral’s native tool calling format. |
|
|
|
For Llama 3.x models configured to output tool calls in JSON format. When using |
|
|
| For InternLM2 models using InternLM’s tool calling format. |
|
|
| For IBM Granite function-calling models. |
|
|
| For Adept Fuyu models. |
|
|
| For Microsoft Phi-3 models configured for JSON output. |
|
|
| For AI21 Labs Jamba models. |
|
Using an incorrect parser can result in runtime errors such as the following:
- Failed tool call extraction
- Malformed tool call requests
- Errors during inference
- Unexpected model behavior
Always verify that the parser matches the expected tool calling format for the model.