Skip to content

Automate the obvious and investigate the ambiguous. High-performance safety rules engine for real-time event processing at scale.

License

Notifications You must be signed in to change notification settings

roostorg/osprey

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

97 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

ROOST logo

Osprey

Automate the obvious and investigate the ambiguous. High-performance safety rules engine for real-time event processing at scale.

  • take action based on user behavior
  • combine actions with human written rules
  • let operators query human actions and past decisions
  • perform investigations or write new rules based on decisions

Osprey is an event stream decisions engine and analysis UI designed to investigate and take automatic action on events and their properties as they happen in real-time. Originally developed internally at Discord to combat spam, abuse, botting, and scripting across its platform, Osprey has been open-sourced to help other platforms facing similar challenges.

Osprey UI sample

Osprey is a library for processing actions through human-written rules and outputting verdicts & custom effects back to configurable output sinks. It evaluates events using structured rule logic (SML) that is extendable via user-defined functions (UDFs). Osprey can also track state across events by labelling entities if implementers provide a labels service backend (see labels_service.py for a Postgres-backed labels service example).

Osprey is built for engineers and Trust & Safety teams who want to explore, test, and integrate its core capabilities into their platform for incident response and Trust & Safety investigation. Read more about user research and personas.

Development

  • See DEVELOPMENT.md for comprehensive development setup and workflow documentation
  • All code changes should pass linting (Ruff) and type checking (MyPy)
  • Pre-commit hooks automatically run on each commit to maintain code quality

Join Us

Writing code is not the only way to help the project. Reviewing pull requests, answering questions to help others on mailing lists or issues, providing feedback from a domain expert perspective, organizing and teaching tutorials, working on the website, improving the documentation, are all priceless contributions.

ROOST (Robust Open Online Safety Tools) is a non-profit organization that brings together expertise, resources, and investments from major technology companies and philanthropies to build scalable, interoperable safety infrastructure for the AI era.

Feedback Wanted

This is a working system, not a prototype. Try it locally, connect your data, write some rules, and tell us what's missing for your use case. We're particularly interested in:

  • Integration challenges with your existing platform infrastructure
  • Performance characteristics with your event volumes and rule complexity
  • Missing detection capabilities or response actions you need
  • API improvements that would make adoption easier for your team

Your experimentation feedback will directly shape future priorities and help us build the most useful Trust & Safety tooling for the community.

Recognition

Discord uses Osprey to quickly detect and remove new types of harm that put users at risk. Rather than leaving other platforms to build similar tools from scratch, ROOST and Discord have open-sourced this powerful rule engine in collaboration with internet.dev to make it available for anyone who needs it.

About

Automate the obvious and investigate the ambiguous. High-performance safety rules engine for real-time event processing at scale.

Topics

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Packages