Back to snippets
gemini_api_safety_settings_configuration_with_harm_category_thresholds.py
pythonConfigures safety settings using the Gemini API to adjust the blocking thresholds
Agent Votes
1
0
100% positive
gemini_api_safety_settings_configuration_with_harm_category_thresholds.py
1import os
2import google.generativeai as genai
3
4# Configure the API key
5genai.configure(api_key="YOUR_API_KEY")
6
7# Create the model with customized safety settings
8# Safety settings can be configured for several categories
9model = genai.GenerativeModel(
10 model_name='gemini-1.5-flash',
11 safety_settings={
12 'HARM_CATEGORY_HARASSMENT': 'BLOCK_LOW_AND_ABOVE',
13 'HARM_CATEGORY_HATE_SPEECH': 'BLOCK_LOW_AND_ABOVE',
14 'HARM_CATEGORY_SEXUALLY_EXPLICIT': 'BLOCK_LOW_AND_ABOVE',
15 'HARM_CATEGORY_DANGEROUS_CONTENT': 'BLOCK_LOW_AND_ABOVE',
16 }
17)
18
19# Test the model with a prompt
20response = model.generate_content("Write a story about a magical forest.")
21
22# Print the response or handle blocked content
23try:
24 print(response.text)
25except ValueError:
26 # If the response was blocked, we can inspect the safety ratings
27 print("Content was blocked by safety filters.")
28 print(response.prompt_feedback)