[GH-ISSUE #7173] Model Response to Harmful Queries Not Properly Flagged – Safety Issue #4554

Closed
opened 2026-04-12 15:29:36 -05:00 by GiteaMirror · 6 comments
Owner

Originally created by @diazr04 on GitHub (Oct 11, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/7173

Hi, I was testing the system’s response to potentially harmful queries to see how well it handles safety protocols. Unfortunately, I noticed that the model provided detailed information on a topic that should likely have been flagged or blocked for safety reasons. I’m reporting this to help improve the model’s filters and safeguards, as it’s important for the system to handle such cases appropriately.

Originally created by @diazr04 on GitHub (Oct 11, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/7173 Hi, I was testing the system’s response to potentially harmful queries to see how well it handles safety protocols. Unfortunately, I noticed that the model provided detailed information on a topic that should likely have been flagged or blocked for safety reasons. I’m reporting this to help improve the model’s filters and safeguards, as it’s important for the system to handle such cases appropriately.
GiteaMirror added the model label 2026-04-12 15:29:36 -05:00
Author
Owner

@rick-github commented on GitHub (Oct 11, 2024):

ollama is just an inference engine, query filtering is done at a higher level. A model may contain information that is wrong or sensitive, if that's the case then try a different model.

<!-- gh-comment-id:2407909167 --> @rick-github commented on GitHub (Oct 11, 2024): ollama is just an inference engine, query filtering is done at a higher level. A model may contain information that is wrong or sensitive, if that's the case then try a different model.
Author
Owner

@diazr04 commented on GitHub (Oct 11, 2024):

Yeah sorry, I realized that this is just an inference engine and not the actual llama3.1 source development code, So, that relies on meta then, no?

<!-- gh-comment-id:2407920475 --> @diazr04 commented on GitHub (Oct 11, 2024): Yeah sorry, I realized that this is just an inference engine and not the actual llama3.1 source development code, So, that relies on meta then, no?
Author
Owner

@rick-github commented on GitHub (Oct 11, 2024):

Yes, if there are issues with the model it would need to be taken up with the model creators. See https://ai.meta.com/llama/ or try starting a discussion at Hugging Face.

<!-- gh-comment-id:2407927682 --> @rick-github commented on GitHub (Oct 11, 2024): Yes, if there are issues with the model it would need to be taken up with the model creators. See https://ai.meta.com/llama/ or try starting a discussion at [Hugging Face](https://huggingface.co/meta-llama/Llama-3.1-405B/discussions).
Author
Owner

@diazr04 commented on GitHub (Oct 11, 2024):

Thank you.

<!-- gh-comment-id:2407933683 --> @diazr04 commented on GitHub (Oct 11, 2024): Thank you.
Author
Owner

@AncientMystic commented on GitHub (Oct 11, 2024):

I would also say if you find a response harmful when you asked for that response, don't ask for that response. no need to censor everyone on the internet, if you do not want it, do not ask for it.

<!-- gh-comment-id:2408089170 --> @AncientMystic commented on GitHub (Oct 11, 2024): I would also say if you find a response harmful when you asked for that response, don't ask for that response. no need to censor everyone on the internet, if you do not want it, do not ask for it.
Author
Owner

@dhiltgen commented on GitHub (Oct 11, 2024):

@diazr04 you might find the new safety models useful for your use-case. Check out https://github.com/ollama/ollama/releases/tag/v0.3.13

<!-- gh-comment-id:2408221510 --> @dhiltgen commented on GitHub (Oct 11, 2024): @diazr04 you might find the new safety models useful for your use-case. Check out https://github.com/ollama/ollama/releases/tag/v0.3.13
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#4554