AI Against Humanity
Back to Geopolitics

Geopolitics Artifacts

2 artifacts

anthropic pentagon ai conflict

Anthropic vs. Pentagon: Legal and Ethical Battles

Updated April 3, 2026 · 5 sources

The ongoing conflict between Anthropic, a prominent AI firm, and the U.S. Department of Defense (DoD) has escalated significantly. The Pentagon has pressured Anthropic for unrestricted access to its AI system, Claude, for military applications, including mass surveillance and autonomous weaponry. Anthropic's CEO, Dario Amodei, has firmly resisted these demands, citing ethical concerns and the potential for misuse of AI technologies. Following a breakdown in negotiations, the Pentagon designated Anthropic as an 'unacceptable risk to national security,' leading to a lawsuit from the company. Recent court rulings have favored Anthropic, halting the Pentagon's actions and questioning the legality of its designation. Meanwhile, the Pentagon is exploring alternative partnerships and considering allowing other AI firms to train on classified data, raising further ethical implications regarding military reliance on AI. The situation has drawn scrutiny from lawmakers and the public, highlighting the critical intersection of technology, ethics, and national security.

Read Artifact
openai gpt5 controversy

OpenAI's GPT-5 Launch: Ethical and Psychological Concerns

Updated March 26, 2026 · 2 sources

The launch of OpenAI's GPT-5 model has ignited significant debate over the ethical implications of advanced AI technologies. While the model enhances speed and accuracy, users have criticized its corporate tone, which detracts from the conversational experience they valued in previous iterations. OpenAI's shift towards product enhancement has led to the departure of key research staff, raising concerns about the future of foundational AI research. The introduction of advertisements in ChatGPT has further fueled fears regarding user privacy and trust, with former employees resigning in protest. Additionally, OpenAI's decision to retire the GPT-4o model has caused distress among users who formed emotional bonds with the AI, leading to lawsuits citing psychological harm. Recent developments, including the launch of GPT-5.4, which enhances autonomous capabilities, have complicated the ethical landscape, particularly in light of OpenAI's military partnerships and the controversial plans for an 'adult mode' that were ultimately shelved due to backlash.

Read Artifact