Anthropic Risk Sentinel
7.2
An AI safety and ethical alignment monitoring platform internally for large language models (LLMs) and externally for investigating misuse cases such as the Gemini incident. Uses advanced sentiment analysis and pattern recognition to detect and flag instances of potentially harmful AI-generated content, proactively identifying and mitigating risks like malicious advice. The core problem directly addresses and seeks to promote
220h
mvp estimate
7.2
viability grade
0
views
technology stack
Python
Difficult
PostgreSQL
Medium
inspired by
Pentagon labels Anthropic a supply chain risk after bullying accusations.