Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Enable potentially abusive user detection insights from calls to Azure endpoint #15

Open
6 tasks
SonOfLope opened this issue Jul 26, 2024 · 0 comments
Open
6 tasks

Comments

@SonOfLope
Copy link
Collaborator

SonOfLope commented Jul 26, 2024

Context

To comply with the Government of Canada's Digital Standards, we are implementing potentially abusive user detection insights in our applications. This initiative focuses on addressing security and privacy risks, designing ethical services, and ensuring responsible data stewardship. The detection system utilizes Azure OpenAI service to analyze user requests and identify potentially harmful content or behaviors. When flagged, these instances are summarized in a report available in the Azure OpenAI Studio, enabling proactive management and a safer user environment.

to discuss

As of now, the requests made from our different applications don't have user information. We would need to implement changes so that calls include the user doing the prompt. We need to evaluate if this is worth considering what the metric actually adds. This is what we would get from adding user information :
image
https://learn.microsoft.com/en-us/azure/ai-services/openai/how-to/risks-safety-monitor#report-description-1

Take note that it doesn't give the prompt to have an idea if the blocking request is a false postive.

Actions to take

  • Deploy a Azure Data Explorer Cluster to collect metrics on potentially abusive usage of models
  • Connect models metrics to Data Explorer Cluster
  • Update Louis deployment to potentially add user metadata to litellm calls to our models to enable user tracking of abusive usage
  • Update Librechat deployment to potentially add user metadata to litellm calls to our models to enable user tracking of abusive usage
  • Update Fertiscan deployment to potentially add user metadata to litellm calls to our models to enable user tracking of abusive usage
  • Update Finesse deployment to potentially add user metadata to litellm calls to our models to enable user tracking of abusive usage
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: Todo
Development

No branches or pull requests

1 participant