Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Benchmarking of gpt-3.5-turbo compliance breach predictions #111

Open
ojus1 opened this issue Jun 1, 2023 · 1 comment
Open

Benchmarking of gpt-3.5-turbo compliance breach predictions #111

ojus1 opened this issue Jun 1, 2023 · 1 comment

Comments

@ojus1
Copy link

ojus1 commented Jun 1, 2023

I've been playing around with ChatGPT (api) with many variations of prompts. It is spitting out complete nonsense reasoning and breach predictions. Has any benchmarking been done for the compliance breach prediction pipeline? If so, I would love to look at the documentation.

@harshithere

@harshithere
Copy link
Contributor

Short answer: Nope, no benchmarking done
The consensus of using this comes from the prototype breach reports that @lepisma was generating last month. The idea was that we want a system that provides some mechanism which is better than the current method (randomly looking through calls or relying on client feedback). Thus large number of FPs is not a concern. Providing baseline accuracy numbers and tuning this system for better performance were kept as future tasks.
Nonetheless, I'm curious what is 'completely nonsense' reasoning. Lets schedule a call for that

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants