Search Results for "ai safety"
US FDA Calls Meeting on Blockchain, AI and the New Era of Smarter Food Safety
The FDA will hold a meeting on October 21st, with the aim of gathering input and feedback to help shape an FDA blueprint for the initiative. The FDA will propose leveraging modern technologies such as blockchain, artificial intelligence, and IoT to improve the transparency and decrease the risks of contamination in the food supply chain.
Guaranteed Safe AI Systems: A Solution for the Future of AI Safety?
Exploring the potential of guaranteed safe AI systems in ensuring the safety and reliability of artificial general intelligence (AGI).
Anthropic Unveils Initiative to Enhance Third-Party AI Model Evaluations
Anthropic announces a new initiative aimed at funding third-party evaluations to better assess AI capabilities and risks, addressing the growing demand in the field.
Anthropic Expands AI Model Safety Bug Bounty Program
Anthropic broadens its AI model safety bug bounty program to address universal jailbreak vulnerabilities, offering rewards up to $15,000.
OpenAI Releases Comprehensive GPT-4o System Card Detailing Safety Measures
OpenAI's report on GPT-4o highlights extensive safety evaluations, red teaming, and risk mitigations prior to release.
Ensuring AI Reliability: NVIDIA NeMo Guardrails Integrates Cleanlab's Trustworthy Language Model
NVIDIA's NeMo Guardrails, in collaboration with Cleanlab's Trustworthy Language Model, aims to enhance AI reliability by preventing hallucinations in AI-generated responses.
NVIDIA NeMo Guardrails Enhance LLM Streaming for Safer AI Interactions
NVIDIA introduces NeMo Guardrails to enhance large language model (LLM) streaming, improving latency and safety for generative AI applications through real-time, token-by-token output validation.
NVIDIA Introduces Safety Measures for Agentic AI Systems
NVIDIA has launched a comprehensive safety recipe to enhance the security and compliance of agentic AI systems, addressing risks such as prompt injection and data leakage.
OpenAI Enhances GPT-5 for Sensitive Conversations with New Safety Measures
OpenAI has released an addendum to the GPT-5 system card, showcasing improvements in handling sensitive conversations with enhanced safety benchmarks.
President Biden Amplifies AI Safety and Security Measures with Executive Order
President Biden has issued an Executive Order on October 30, 2023, aiming to improve AI safety, security, and trustworthiness. The order requires rigorous testing of critical AI systems, advocates for data privacy legislation, and promotes AI's positive impact on healthcare, education, and the labor market.
UK to Host First International AI Safety Conference in November
The United Kingdom is set to host the world's first international conference on AI safety on November 1-2, 2023. The summit aims to position the UK as a mediator in tech discussions between the US, China, and the EU. Prime Minister Rishi Sunak will host the event at Bletchley Park, featuring notable attendees like US Vice President Kamala Harris and Google DeepMind CEO Demis Hassabis. The conference will focus on the existential risks posed by AI, among other safety concerns.
Exploring AI Stability: Navigating Non-Power-Seeking Behavior Across Environments
The research explores AI's stability in non-power-seeking behaviors, revealing that certain policies maintain non-resistance to shutdown across similar environments, providing insights into mitigating risks associated with power-seeking AI.