nymnympseudonym

joined 1 day ago
[–] nymnympseudonym@piefed.social 1 points 20 hours ago* (last edited 20 hours ago)

Perplexity prompt: "quotes from openai about safety testing for gpt-oss"

Reply below:


Safety Testing Measures

  • OpenAI’s safety process for open-weight release included technical audits (robustness, bias detection), structured adversarial testing, and evaluation of prompt-injection and jailbreak vulnerabilities.
  • The organization also employed both automated and human review to test for outputs that could cause harm, leveraging cross-disciplinary teams including outside researchers, ethicists, and cybersecurity experts.
  • Feedback from these rounds led to incremental model improvements before the open weights went public.

Transparency and External Collaboration

  • OpenAI has collaborated with third-party security and ethics researchers to validate its safety protocols and stress-test new models prior to release.
  • The company acknowledges that “Releasing open weights is a significant responsibility due to risks of misuse. We want to be transparent about our process and invite the AI community to help report and address issues that may arise post-release.”
[–] nymnympseudonym@piefed.social 7 points 22 hours ago (4 children)

maybe our averages are different

[–] nymnympseudonym@piefed.social 9 points 23 hours ago* (last edited 23 hours ago) (6 children)

As a person who has been managing software development teams for 30+ years, I have an observation.
Invariably, some employees are "average". Not super geniuses, not workaholics, but people who (say) have been doing a good job with customer support. Generally they can code simple things and know the OS versions we support as a power user -- but not as well as a sysadmin.

I do find that if I tell them to use ChatGPT to help debug issues, they do almost as well as if a sysadmin or more experienced programmer had picked up the ticket. It gets better troubleshooting, they maybe fix an actual root cause bug in our product code.

[–] nymnympseudonym@piefed.social 1 points 1 day ago (1 children)

Whatever it is, it's probably in the Epstein papers