[GH-ISSUE #12288] deepseek-v3.1: upstream error on tool parameters "required": null #8168

Open
opened 2026-04-12 20:35:04 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @matscalia on GitHub (Sep 14, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12288

Originally assigned to: @ParthSareen on GitHub.

What is the issue?

Minimal repro (tested on turbo):

#!/usr/bin/bash
curl https://ollama.com/api/chat \
  -X POST \
  -H "Authorization: $OLLAMA_TURBO_KEY" \
  -d '{
        "model": "deepseek-v3.1:671b",
        "messages": [
          {
            "role": "user",
            "content": "What is the weather like in London? What are the conditions in Toronto?"
          }
        ],
        "tools": [
          {
            "type": "function",
            "function": {
              "name": "get_weather_conditions",
              "description": "Get the weather conditions for a city",
              "parameters": {
                "type": "object",
                "required": null,
                "properties": {
                  "city": {
                    "type": "string",
                    "description": "The name of the city"
                  }
                }
              }
            }
          }
        ]
      }'

the same command works well for gpt-oss models and no upstream error is returned, that is, the expected behavior.

Relevant log output


OS

No response

GPU

No response

CPU

No response

Ollama version

Turbo

Originally created by @matscalia on GitHub (Sep 14, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12288 Originally assigned to: @ParthSareen on GitHub. ### What is the issue? Minimal repro (tested on turbo): ```bash #!/usr/bin/bash curl https://ollama.com/api/chat \ -X POST \ -H "Authorization: $OLLAMA_TURBO_KEY" \ -d '{ "model": "deepseek-v3.1:671b", "messages": [ { "role": "user", "content": "What is the weather like in London? What are the conditions in Toronto?" } ], "tools": [ { "type": "function", "function": { "name": "get_weather_conditions", "description": "Get the weather conditions for a city", "parameters": { "type": "object", "required": null, "properties": { "city": { "type": "string", "description": "The name of the city" } } } } } ] }' ``` the same command works well for gpt-oss models and no upstream error is returned, that is, the expected behavior. ### Relevant log output ```shell ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version Turbo
GiteaMirror added the cloudbug labels 2026-04-12 20:35:04 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#8168