OpenAI is moving to publish the results of its internal AI model safety evaluations more regularly in what the outfit is pitching as an effort to increase transparency. On Wednesday, OpenAI launched the Safety Evaluations Hub, a webpage showing how the company’s models score on various tests for harmful content generation, jailbreaks, and hallucinations. OpenAI…
Category: Safety
AI, Global Security News, gpt 4.1, openai, Safety, Security
OpenAI’s GPT-4.1 may be less aligned than the company’s previous AI models
In mid-April, OpenAI launched a powerful new AI model, GPT-4.1, that the company claimed “excelled” at following instructions. But the results of several independent tests suggest the model is less aligned — that is to say, less reliable — than previous OpenAI releases. When OpenAI launches a new model, it typically publishes a detailed technical…
AI, Global Security News, openai, Safety
OpenAI may ‘adjust’ its safeguards if rivals release ‘high-risk’ AI
In an update to its Preparedness Framework, the internal framework OpenAI uses to decide whether AI models are safe and what safeguards, if any, are needed during development and release, OpenAI said that it may “adjust” its requirements if a rival AI lab releases a “high-risk” system without comparable safeguards. The change reflects the increasing…
agility robotics, AI, figure, Global IT News, Global Security News, Hardware, humanoid, humanoid robot, OSHA, robotics, Safety
Figure AI details plan to improve humanoid robot safety in the workplace
Safety is often overlooked in the rush to bring humanoid robots to the workplace. As high-profile corporations ranging from retailers like Amazon to carmakers such as Mercedes and BMW have announced humanoid pilots for factories and warehouses, conversations around worker safety are regularly buried beneath industry hype. One Bay Area-robotics firm is hoping to correct…
AI, AI chatbots, character.ai, discord, generative ai, Global IT News, Global Security News, Government & Policy, privacy, Safety, Social
Texas AG is investigating Character.AI, other platforms over child safety concerns
Texas Attorney General Ken Paxton on Thursday launched an investigation into Character.AI and 14 other technology platforms over child privacy and safety concerns. The investigation will assess whether Character.AI — and other platforms that are popular with young people, including Reddit, Instagram and Discord — conform to Texas’ child privacy and safety laws. The investigation…