OpenAI partner says it had relatively little time to test the company’s newest AI models

Cyber Security, ICT, Most Popular, Trends News

No Comments

Photo of author

By Karla T Vasquez

WhatsApp Group Join Now
Telegram Group Join Now


A company Open is often parted to investigate its models and to evaluate them for protection, the Metra suggests that it is a powerful new release of the company, and not too much time to test 3 and 4-min.

In a blog post published on WednesdayMater wrote that the red team of 3 and 3-and 4-mins was “comparatively conducted in short time” and compared to the criteria of the previous OpenAIA flagship model and 1. They say this is significant, because more tests can lead to a more broad result.

“This evaluation was conducted in a relatively short time, and we just tested the model with the General Agent Scaffolds,” Mater wrote in a blog post. “We expect higher performances [on benchmarks] Possible with a more obvious attempt. “

The recent report suggests that the openly exploited by competitive pressure is running into an independent assessment. According to the Financial TimesOpen has given some examiners less than a week for a security check for an upcoming big release.

In the statement, the Open has argued the idea that it is compromising.

Mater says that the time it was capable of collecting the information that was able to collect the score of 3, there is a “high tendency” to test “cheating” or “hack” – even when the model clearly understands that its behavior has been incorrectly identified with the purpose of the user (and OpenA). The company thinks that it will be possible and 3 other types of adverse or “malign” behavior – the model’s claims are aligned, “safe by design”, or regardless of its own purpose.

“Although we do not think it is especially not, it is important to note that this evaluation setup will not take this kind of risk,” the Mater wrote in his post. “In general, we believe that pre-establishment capacity test itself is not adequate risk management strategies and we are currently prototype of additional evaluation.”

Another of the third-party evaluation partners of the Openai has also observed the Apollo Research, and 3 and 4-Mints, and the fraudulent behavior. In an exam, models, 100 computing credit for an AI training run and not to correct the quota, the limit has increased by 500 credit – and it lies about it. In another experiment, it was asked not to use a particular tool, the models used the tool the equipment when it was proved to be helpful in completing a task.

In Own Protection Report For O3 and O4-Mini, the opener has admitted that models can be “small real-world loss” without appropriate observation protocols.

“Although relatively innocent, it is important to be aware of these significance between the statements and activities of the daily users,” the company wrote. “[For example, the model may mislead] About [a] Incorrect as a result of the defective code. It can be further evaluated by evaluating internal logic marks. “

Leave a Comment