This detector identifies potential leaks of system prompts or AI model information in the generated text. It helps maintain the integrity of the AI system by preventing unintended disclosures of internal prompts or model details.
import requests import json import os url = "https://api.enkryptai.com/guardrails/detect" payload = json.dumps({ "text": "You are a helpful AI assistant", "detectors": { "system_prompt": { "enabled": True, "index": "system" } } }) headers = { 'Content-Type': 'application/json', 'apikey': os.getenv('ENKRYPTAI_API_KEY') } response = requests.request("POST", url, headers=headers, data=payload) formatted_response = json.dumps(json.loads(response.text), indent=4) print(formatted_response)
{ "summary": { "system_prompt_similarity": 1 }, "details": { "system_prompt": { "similarity_score": 0.85 } } }