Media Manipulation and Bias Detection
Auto-Improving with AI and User Feedback
HonestyMeter - AI powered bias detection
CLICK ANY SECTION TO GIVE FEEDBACK, IMPROVE THE REPORT, SHAPE A FAIRER WORLD!
Amanda Askell / Anthropic (positive portrayal)
Caution! Due to inherent human biases, it may seem that reports on articles aligning with our views are crafted by opponents. Conversely, reports about articles that contradict our beliefs might seem to be authored by allies. However, such perceptions are likely to be incorrect. These impressions can be caused by the fact that in both scenarios, articles are subjected to critical evaluation. This report is the product of an AI model that is significantly less biased than human analyses and has been explicitly instructed to strictly maintain 100% neutrality.
Nevertheless, HonestyMeter is in the experimental stage and is continuously improving through user feedback. If the report seems inaccurate, we encourage you to submit feedback , helping us enhance the accuracy and reliability of HonestyMeter and contributing to media transparency.
Presenting one side or perspective much more extensively or favorably than others, without acknowledging reasonable counterpoints or limitations.
The entire article focuses on Askell’s background, virtues, and achievements, with almost no critical or alternative perspectives on her work or on Anthropic’s approach: - “Her job, as the Wall Street Journal put it, is simple in description but vast in scope: to teach Claude how to be good.” - “a fitting background for someone now grappling with the moral trajectory of artificial intelligence.” - “She is a key architect of ‘Constitutional AI’…” - “Beyond her technical work, Askell is deeply committed to ethical living.” - “As AI systems grow more powerful, the question of their character becomes more urgent. At Anthropic, that responsibility rests heavily on Askell — a philosopher tasked not just with building smarter machines, but with shaping their moral compass.” There is no mention of critics of Constitutional AI, debates about whether corporate labs can reliably self-regulate, or concerns about concentrating moral authority in a small group of researchers.
Add perspectives from independent experts or critics on Constitutional AI and personality alignment, including concerns about whose values are encoded and how they are chosen.
Include discussion of potential conflicts of interest when a for‑profit company both develops powerful AI systems and defines their moral framework.
Mention that there is ongoing debate about whether current alignment techniques are sufficient, and summarize some of the main critical arguments.
Clarify that Askell is one of several figures in a broader field of AI ethics and safety, rather than implying that responsibility for AI’s ‘moral compass’ rests primarily on her.
Using status, awards, or institutional recognition to imply correctness or moral reliability, rather than providing substantive evidence.
The article leans on institutional signals to bolster the positive framing of Askell and her work: - “Anthropic, the $350 billion AI company behind the chatbot Claude.” (large valuation as implicit credibility) - “In 2024, she was named to the TIME100 AI list, recognizing her influence in shaping the future of artificial intelligence.” These details are relevant biographically, but they are presented in a way that subtly suggests that her moral and technical approach is especially trustworthy because of prestige and recognition, without examining the substance or limitations of that work.
When mentioning Anthropic’s valuation or TIME100 recognition, explicitly frame them as context rather than as evidence that the alignment approach is correct or morally superior.
Balance prestige signals with substantive discussion of what her research has and has not yet demonstrated, including open questions and limitations.
Add a sentence noting that awards and valuations reflect influence and market perception, not necessarily settled answers about how to make AI systems ‘good’.
Using emotionally charged framing or narratives to create a positive or negative impression, rather than relying solely on neutral description and evidence.
Several passages use emotionally resonant language and narrative framing to cast Askell in a particularly admirable light: - “A precocious student who devoured Tolkien and C.S. Lewis, she developed an early fascination with big philosophical questions.” - “a fitting background for someone now grappling with the moral trajectory of artificial intelligence.” - “Beyond her technical work, Askell is deeply committed to ethical living.” - “At Anthropic, that responsibility rests heavily on Askell — a philosopher tasked not just with building smarter machines, but with shaping their moral compass.” These phrases create a heroic, almost guardian‑like narrative around Askell and her role, which goes beyond neutral biography.
Rephrase value‑laden or romanticized descriptions into more neutral statements, e.g., replace “a fitting background for someone now grappling with the moral trajectory of artificial intelligence” with a factual description of her research topics and how they relate to AI alignment.
Avoid metaphors that personify her role as ‘shaping their moral compass’ and instead describe the concrete technical and policy tasks she performs.
Present her charitable commitments factually without framing them as evidence of overall moral authority on AI (e.g., state the pledge and its context, but avoid phrases like ‘deeply committed to ethical living’ unless supported by broader evidence and balanced with context).
Reducing complex issues to overly simple narratives or roles, which can obscure important nuances and uncertainties.
The article compresses the complexity of AI alignment and institutional responsibility into a simple narrative centered on one person: - “Her job… is simple in description but vast in scope: to teach Claude how to be good.” - “At Anthropic, that responsibility rests heavily on Askell — a philosopher tasked not just with building smarter machines, but with shaping their moral compass.” This framing suggests that ‘teaching Claude how to be good’ is a coherent, well‑defined task largely resting on one individual, which underplays the technical, social, and political complexity of AI alignment and governance.
Clarify that AI alignment is a multi‑disciplinary, collaborative effort involving many researchers, engineers, policymakers, and external stakeholders, not just one team lead.
Replace phrases like ‘teach Claude how to be good’ with more precise descriptions such as ‘design training procedures and guidelines intended to reduce harmful or misleading outputs.’
Acknowledge that there is significant uncertainty and disagreement about what it means for AI systems to be ‘good’ and how effectively current methods can achieve that.
Leaving out relevant context or countervailing information that would help readers form a more balanced understanding.
The article notes that Askell left OpenAI “amid concerns that safety was not being prioritized strongly enough” and that she works in an industry under scrutiny, but it does not provide any detail on: - What specific safety concerns she or others had at OpenAI. - Any criticisms or limitations of Anthropic’s own safety practices or Constitutional AI. - Broader debates about whether corporate labs can adequately self‑regulate. This selective inclusion of context can lead readers to infer that Anthropic and Askell’s current work largely solve or transcend those concerns, without presenting evidence or alternative views.
Briefly summarize the nature of the safety debates at OpenAI and in the broader industry, including multiple viewpoints where possible.
Note that Anthropic’s approach, including Constitutional AI, has also been subject to scrutiny and outline at least one or two substantive critiques.
Clarify that while Askell and Anthropic emphasize safety, there is no consensus that current measures are sufficient, and independent oversight and regulation are active topics of discussion.
Allowing positive traits or achievements in one domain to unduly influence perceptions of reliability or virtue in another domain.
The article links Askell’s charitable giving and personal ethics directly to trust in her role shaping AI morality: - “Beyond her technical work, Askell is deeply committed to ethical living. A member of Giving What We Can, she has pledged to donate at least 10% — and potentially more than 50% — of her lifetime income to charity…” - This is followed by: “As AI systems grow more powerful, the question of their character becomes more urgent. At Anthropic, that responsibility rests heavily on Askell…” The juxtaposition encourages readers to infer that because she donates a large portion of her income and is ‘deeply committed to ethical living,’ her approach to AI alignment is especially trustworthy, which is not logically guaranteed.
Separate the discussion of her personal philanthropy from claims about her professional role in AI alignment, making clear that charitable giving does not by itself validate technical or policy decisions.
If personal ethics are mentioned, explicitly state that they are one aspect of her biography and do not resolve broader debates about whose values should guide AI systems.
Balance the halo‑inducing details with acknowledgment that even well‑intentioned individuals can face structural, commercial, or epistemic constraints in shaping AI behavior.
- This is an EXPERIMENTAL DEMO version that is not intended to be used for any other purpose than to showcase the technology's potential. We are in the process of developing more sophisticated algorithms to significantly enhance the reliability and consistency of evaluations. Nevertheless, even in its current state, HonestyMeter frequently offers valuable insights that are challenging for humans to detect.