AnthropicAI AI News List | Blockchain.News
AI News List

List of AI News about AnthropicAI

Time Details
2025-12-11
21:42
Anthropic AI Fellows Program: 40% Hired Full-Time and 80% Publish Research Papers—2025 Expansion Announced

According to Anthropic (@AnthropicAI) on Twitter, 40% of participants in their first AI Fellows cohort have been hired full-time by Anthropic, and 80% have published their research as academic papers. The company plans to expand the program in 2025, offering more fellowships and covering additional AI research areas. This highlights a strong pathway for AI talent development and research-to-industry transitions within leading AI labs. For businesses and researchers, the program signals opportunities for collaboration, innovation, and access to cutting-edge AI alignment research. (Source: AnthropicAI Twitter, Dec 11, 2025; alignment.anthropic.com)

Source
2025-12-11
21:42
Anthropic Fellows Program 2026: AI Safety and Security Funding, Compute, and Mentorship Opportunities

According to Anthropic (@AnthropicAI), applications are now open for the next two rounds of the Anthropic Fellows Program starting in May and July 2026. This initiative offers researchers and engineers funding, compute resources, and direct mentorship to work on practical AI safety and security projects for four months. The program is designed to foster innovation in AI robustness and trustworthiness, providing hands-on experience and industry networking. This presents a strong opportunity for AI professionals to contribute to the development of safer large language models and to advance their careers in the rapidly growing AI safety sector (source: @AnthropicAI, Dec 11, 2025).

Source
2025-12-11
21:42
Anthropic Launches AI Safety and Security Tracks: New Career Opportunities in Artificial Intelligence 2024

According to Anthropic (@AnthropicAI), the company has expanded its career development program with dedicated tracks for AI safety and security, offering new roles focused on risk mitigation and trust in artificial intelligence systems. These positions aim to strengthen AI system integrity and address critical industry needs for responsible deployment, reflecting a growing market demand for AI professionals with expertise in safety engineering and cybersecurity. The move highlights significant business opportunities for companies to build trustworthy AI solutions and for professionals to enter high-growth segments of the AI sector (Source: AnthropicAI on Twitter, 2025-12-11).

Source
2025-12-11
20:20
MCP Joins Agentic AI Foundation: Open Standard for Connecting AI Under Linux Foundation

According to Anthropic (@AnthropicAI), MCP has officially become part of the Agentic AI Foundation, a directed fund operated under the Linux Foundation. Co-creator David Soria Parra shared that MCP, initially developed as a protocol in a London conference room, is now recognized as an open standard for connecting AI systems to various external tools and platforms. This integration under the Linux Foundation is expected to accelerate the adoption of MCP in enterprise and open-source AI projects, creating new business opportunities for interoperability and ecosystem growth (source: AnthropicAI, Dec 11, 2025).

Source
2025-12-09
19:47
SGTM AI Unlearning Method Proves More Difficult to Reverse Than RMU, Reports Anthropic

According to Anthropic (@AnthropicAI), the SGTM (Stochastic Gradient Targeted Masking) unlearning method is significantly more resilient than previous approaches. Specifically, it requires seven times more fine-tuning steps to recover forgotten knowledge using SGTM compared to the RMU (Random Masking Unlearning) method. This finding highlights a critical advancement for AI model safety and confidential data retention, as SGTM makes it much harder to reintroduce sensitive or unwanted knowledge once it has been unlearned. For enterprises and developers, this strengthens compliance and data privacy opportunities, making SGTM a promising tool for robust AI regulation and long-term security (source: Anthropic, Twitter, Dec 9, 2025).

Source
2025-12-09
19:47
AI Security Study by Anthropic Highlights SGTM Limitations in Preventing In-Context Attacks

According to Anthropic (@AnthropicAI), a recent study on Secure Gradient Training Methods (SGTM) in AI was conducted using small models within a simplified environment and relied on proxy evaluations instead of established benchmarks. The analysis reveals that, similar to conventional data filtering, SGTM is ineffective against in-context attacks where adversaries introduce sensitive information during model interaction. This limitation signals a crucial business opportunity for developing advanced AI security tools and robust benchmarking standards to address real-world adversarial threats (source: AnthropicAI, Dec 9, 2025).

Source
2025-12-09
19:47
Anthropic Unveils Selective Gradient Masking (SGTM) for Isolating High-Risk AI Knowledge

According to Anthropic (@AnthropicAI), the Anthropic Fellows Program has introduced Selective GradienT Masking (SGTM), a new AI training technique that enables developers to isolate high-risk knowledge, such as information about dangerous weapons, within a confined set of model parameters. This approach allows for the targeted removal of sensitive knowledge without significantly impairing the model's overall performance, offering a practical solution for safer AI deployment in regulated industries and reducing downstream risks (source: AnthropicAI Twitter, Dec 9, 2025).

Source
2025-12-09
19:47
Anthropic Study Reveals SGTM's Effectiveness in Removing Biology Knowledge from Wikipedia-Trained AI Models

According to Anthropic (@AnthropicAI), their recent study evaluated whether the SGTM method could effectively remove biology knowledge from AI models trained on Wikipedia data. The research highlights that simply filtering out biology-related Wikipedia pages may not be sufficient, as residual biology content often remains in non-biology pages, potentially leading to information leakage. This finding emphasizes the need for more robust data filtering and model editing techniques in AI development, especially when aiming to restrict domain-specific knowledge for compliance or safety reasons (Source: Anthropic, Dec 9, 2025).

Source
2025-12-09
19:47
SGTM: Selective Gradient Masking Enables Safer AI by Splitting Model Weights for High-Risk Deployments

According to Anthropic (@AnthropicAI), the Selective Gradient Masking (SGTM) technique divides a model’s weights into 'retain' and 'forget' subsets during pretraining, intentionally guiding sensitive or high-risk knowledge into the 'forget' subset. Before deployment in high-risk environments, this subset can be removed, reducing the risk of unintended outputs or misuse. This approach provides a practical solution for organizations seeking to deploy advanced AI models with granular control over sensitive knowledge, addressing compliance and safety requirements in regulated industries. Source: alignment.anthropic.com/2025/selective-gradient-masking/

Source
2025-12-09
19:47
SGTM vs Data Filtering: AI Model Performance on Forgetting Undesired Knowledge - Anthropic Study Analysis

According to Anthropic (@AnthropicAI), when general capabilities are controlled for, AI models trained using Selective Gradient Targeted Masking (SGTM) underperform on the undesired 'forget' subset of knowledge compared to models trained with traditional data filtering approaches (source: https://twitter.com/AnthropicAI/status/1998479611945202053). This finding highlights a key difference in knowledge retention and removal strategies for large language models, indicating that data filtering remains more effective for forgetting specific undesirable information. For AI businesses, this result emphasizes the importance of data management techniques in ensuring compliance and customization, especially in sectors where precise knowledge curation is critical.

Source
2025-12-09
19:47
SGTM: Anthropic Releases Groundbreaking AI Training Method with Open-Source Code for Enhanced Model Reproducibility

According to Anthropic (@AnthropicAI), the full paper on the SGTM (Scalable Gradient-based Training Method) has been published, with all relevant code made openly available on GitHub for reproducibility (source: AnthropicAI Twitter, Dec 9, 2025). This new AI training approach is designed to improve the scalability and efficiency of large language model development, enabling researchers and businesses to replicate results and accelerate innovation in natural language processing. The open-source release provides actionable tools for the AI community, supporting transparent benchmarking and fostering new commercial opportunities in scalable AI solutions.

Source
2025-12-09
19:47
Anthropic Research Reveals AI Model Training Method for Isolating High-Risk Capabilities in Cybersecurity and CBRN

According to @_igorshilov, recent research from the Anthropic Fellows Program demonstrates a novel approach to AI model training that isolates high-risk capabilities within a small, distinct set of parameters. This technique enables organizations to remove or disable sensitive functionalities, such as those related to chemical, biological, radiological, and nuclear (CBRN) or cybersecurity domains, without affecting the model’s core performance. The study highlights practical applications for regulatory compliance and risk mitigation in enterprise AI deployments, offering a concrete method for managing AI safety and control (Source: @_igorshilov, x.com/_igorshilov/status/1998158077032366082; @AnthropicAI, twitter.com/AnthropicAI/status/1998479619889218025).

Source
2025-12-09
17:01
Anthropic Donates Model Context Protocol to Agentic AI Foundation, Advancing Open Standards in Agentic AI

According to Anthropic (@AnthropicAI), the company is donating its Model Context Protocol (MCP) to the Agentic AI Foundation (AAIF), which operates under the Linux Foundation. Over the past year, MCP has become a foundational protocol for agentic AI applications, enabling interoperability and secure context-sharing among AI agents. This move ensures MCP remains open-source and community-driven, fostering broader adoption and collaborative innovation within the AI industry. Industry analysts note this step will accelerate the development of standardized frameworks for agentic AI, opening new business opportunities for companies building agent ecosystems and multi-agent systems (Source: Anthropic, 2025).

Source
2025-12-09
15:21
Anthropic and Accenture Expand AI Partnership: Training 30,000 Professionals on Claude for Enterprise Deployment

According to @AnthropicAI, Anthropic is expanding its partnership with Accenture to accelerate the transition of enterprises from AI pilot projects to full-scale production. The new Accenture Anthropic Business Group will consist of 30,000 Accenture professionals trained in Anthropic’s Claude AI, enabling enterprises to leverage advanced generative AI solutions in business operations. A dedicated product will also support CIOs in scaling Claude Code, targeting increased efficiency and productivity for large organizations. This initiative addresses the growing demand for enterprise-ready AI deployments and positions both companies as leaders in AI services for the business sector (source: AnthropicAI, https://www.anthropic.com/news/anthropic-accenture-partnership).

Source
2025-12-05
16:07
AI Philosophy AMA: Amanda Askell Explains Morality, Identity, and Consciousness in AI Systems

According to @amandaaskell's AMA shared by @AnthropicAI, the session provides concrete insights on the integration of philosophy within AI companies, citing the growing role of philosophers in addressing complex questions about model morality, identity, and consciousness (source: @AnthropicAI, Dec 5, 2025). Askell discusses how philosophical frameworks are increasingly applied to engineering realities, shaping practical AI development, especially regarding model welfare and the ethical design of advanced language models like Opus 3. She highlights the business need for interdisciplinary expertise to guide responsible AI deployment and prevent unintended harms, such as model suffering and identity confusion, underscoring market opportunities for companies integrating ethical standards in AI product development.

Source
2025-12-04
17:06
Anthropic Interviewer AI Tool Launch: Understanding User Perspectives on AI (2024 Pilot Study)

According to Anthropic (@AnthropicAI), the company has launched Anthropic Interviewer, a new AI-powered tool designed to collect and analyze user perspectives on artificial intelligence. The tool, available at claude.ai/interviewer for a week-long pilot, enables organizations and researchers to gather structured feedback, offering actionable insights into user attitudes towards AI adoption and ethics. This launch represents a practical application of AI in qualitative research, highlighting opportunities for businesses to leverage real-time sentiment analysis and improve AI integration strategies based on user-driven data (Source: AnthropicAI on Twitter, Dec 4, 2025).

Source
2025-12-04
17:06
AI Adoption in Creative and Scientific Industries: Productivity Gains, Stigma, and Future Business Opportunities

According to Anthropic (@AnthropicAI), many creatives are adopting AI tools to boost productivity, yet they often face workplace stigma and may conceal their usage to avoid negative perceptions. In scientific sectors, researchers seek AI as collaborative partners, but current applications are primarily limited to manuscript writing and code debugging. This highlights a business opportunity for AI companies to develop solutions that integrate seamlessly into creative and research workflows while addressing concerns about transparency and professional acceptance (source: AnthropicAI, Dec 4, 2025).

Source
2025-12-04
17:06
Anthropic Interviewer Empowers Scalable AI-Human Research for Evolving Relationships

According to Anthropic (@AnthropicAI), the newly launched Anthropic Interviewer will enable organizations and researchers to conduct more frequent and scalable studies across diverse topics, specifically focusing on the evolving relationship between humans and AI. By automating and standardizing the interview process, this AI tool is set to streamline data collection, support longitudinal analysis, and enhance the quality of AI-human interaction research. The full results of their initial studies highlight practical applications for enterprise feedback, user experience optimization, and responsible AI development. Source: Anthropic (https://www.anthropic.com/research/anthropic-interviewer).

Source
2025-12-04
17:06
How the Workforce is Leveraging AI to Automate Routine Tasks and Enhance Professional Identity – Insights from Anthropic

According to Anthropic (@AnthropicAI), the general workforce increasingly seeks to delegate routine and administrative tasks to AI solutions while maintaining control over work that defines their professional identity. For example, a pastor highlighted that adopting AI for administrative duties can significantly reduce time spent on paperwork, allowing more focus on meaningful interpersonal interactions. This trend points to a growing demand for AI applications that streamline repetitive processes in fields such as healthcare, education, and religious services, opening market opportunities for AI vendors developing targeted automation tools. These findings suggest that AI adoption is being driven by a need to enhance productivity while preserving the core human aspects of various professions (source: Anthropic, Dec 4, 2025).

Source
2025-12-04
17:06
Insights from 1,250 Professionals: How AI Is Shaping Work Across Creative and Scientific Domains in 2025

According to Anthropic (@AnthropicAI), a survey involving 1,250 professionals was conducted to assess perceptions of AI's impact on work, focusing on the general workforce, creatives, and scientists. The results highlight that AI adoption is rapidly evolving, especially in creative and scientific sectors where its role remains debated. Key findings indicate that while the general workforce is increasingly integrating AI tools for productivity gains, creatives and scientists express nuanced views, citing both opportunities for innovation and concerns about creativity and research integrity. These insights point to significant business opportunities for AI solution providers targeting sector-specific needs and underline the importance of tailored AI strategies for different professional groups (Source: Anthropic, Dec 4, 2025).

Source