OpenAI says that for prompts related to biological and chemical threats, it has deployed its latest AI rational models and a new system to monitor 3 and 4-min. The goal of the system is to prevent models from giving suggestions that may instruct someone to carry out a potentially harmful attack, According to the OPENAIThe
The agency says the 3 and the 4-Mint AAII represents a meaningful power growth compared to the previous models of OpenAI, and thus creating new risks in the hands of bad actors. According to the internal criteria of the Openai, 3, especially in the way of creating a certain type of biological threats, is more efficient in answering questions. This reasons-and to reduce other risks-Opnai has created a new monitoring system, describing the company as a “protection-centric argument monitor”.
The contents of the OPENY are rational custom-educated monitors about the policies and 3 and the 4-min-min. It is designed to detect prompts related to biological and chemical risk and to indicate models to refuse to advice on those topics.
For setting up a baseline, Red Timers and 3 and 4-Mins in OpenAE spent about 1,000 hours flagging the Biorisk conversation from 4-min. According to the Openai, during an experiment that imitates the “blocking logic” of its protection monitor, models refused to respond to 98.7% of the time, according to Openai.
Opena acknowledges that its test has not been an account for people who can try new prompts after the monitor is blocked, which is why the company says it will depend on human observation.
According to the agency, the “high risk” of the 3 and 4-Minit Openai does not exceed the threshold. However, the OpenAI, compared to the O 1 and GPT -4, says that the primary versions of 3 and 4 -min -minimum proved to be more helpful in answering the question surrounding biological weapons development.

Openai’s recently updated update is actively tracking how its models can make its models easier to develop chemical and biological threats to corrupt users Preparation structureThe
OpenAI depends on increasingly automatic systems to reduce risk from its models. For example, to prevent GPT -4O’s native image generator from creating elements of child sexual abuse (CSAM) (OpenAI says that it also uses an argument similar to the company deployed for 3 and 4-min.
Nevertheless, several researchers have raised concerns that Open is not giving priority to security as much as he should. One of the company’s Red-Timing Partners, Mater, says that there was a relatively short time to test a standard and 3 for fraudulent behavior. Meanwhile, the Open has decided not to publish a security report for its GPT -1.3 model, which was launched this week.
