Close Menu
Gossips Today
  • Tech & Innovation
  • Healthcare
  • Personal Finance
  • Lifestyle
  • Travel
  • Business
  • Recipes
What's Hot

Why Silicon Valley is really talking about fleeing California (it’s not the 5%)

Kaiser affiliates to pay $556M to resolve Medicare Advantage fraud allegations

I'm a Flight Attendant, and This Carry-on From Amazon Is My Secret to Fitting a Month's Worth of Clothes in 1 Bag

Facebook X (Twitter) Instagram
Sunday, January 18
Gossips Today
Facebook X (Twitter) Instagram
  • Tech & Innovation

    Why Silicon Valley is really talking about fleeing California (it’s not the 5%)

    January 18, 2026

    Who gets to inherit the stars? A space ethicist on what we’re not talking about

    January 18, 2026

    Musk wants up to $134B in OpenAI lawsuit, despite $700B fortune

    January 17, 2026

    AI cloud startup Runpod hits $120M in ARR — and it started with a Reddit post  

    January 17, 2026

    Anthropic taps former Microsoft India MD to lead Bengaluru expansion

    January 16, 2026
  • Healthcare

    Kaiser affiliates to pay $556M to resolve Medicare Advantage fraud allegations

    January 18, 2026

    MedPAC steps away from advocating doctor pay be tied to inflation

    January 17, 2026

    HCA names new chief nurse executive

    January 17, 2026

    Medicare Advantage overpayments will total $76B this year: MedPAC

    January 16, 2026

    Trump unveils healthcare affordability plan

    January 16, 2026
  • Personal Finance

    How to Stop Living Paycheck to Paycheck

    September 10, 2025

    Real Estate Report 2024 – Ramsey

    September 9, 2025

    How Much Car Can I Afford?

    September 9, 2025

    21 Cheap Beach Vacations for 2025

    August 5, 2025

    Car Depreciation: How Much Is Your Car Worth?

    August 4, 2025
  • Lifestyle

    Begin Again: How To Finally Find Time For What Matters With Backwards Planning

    January 13, 2026

    It’s Time to Begin Again: 3 Uncomfortable Frameworks That Will Make Your New Year More Meaningful [Audio Essay + Article]

    January 10, 2026

    The Getup: The Winter Visit Outfit

    January 5, 2026

    Free Printable Year End Review Journal: An Easy, Structured Way to Reflect Then Build the New Year

    December 30, 2025

    The Smart Man’s Guide to Winter Style: 26 On-Sale Staples That Do the Heavy Lifting (limited time)

    December 16, 2025
  • Travel

    I'm a Flight Attendant, and This Carry-on From Amazon Is My Secret to Fitting a Month's Worth of Clothes in 1 Bag

    January 18, 2026

    This Over 300-mile U.S. Road Trip Is Called the 'Death Drive'—and It Passes Ghost Towns and a Stunning National Park

    January 17, 2026

    I Was a Gate Agent for Years—Here’s What Most Travelers Get Wrong When Their Flight Is Delayed

    January 17, 2026

    I Spent a Cozy Night in a ‘Literary Oasis’ Above a Nantucket Bookstore—Here’s What It Was Like

    January 16, 2026

    A Seasoned Flight Attendant Revealed the 14 Amazon Must-haves They Always Pack for Flights—Starting at Just $3

    January 16, 2026
  • Business

    This common security measure is draining your workforce

    January 18, 2026

    You’re banned from blocking Trump’s face on your national park pass—but there’s a work-around

    January 17, 2026

    FDA commissioner’s drug review plan sparks alarm across the agency

    January 17, 2026

    Australia’s social media ban for children has already wiped out 4.7 million accounts

    January 16, 2026

    The U.S. is suspending immigrant visas from 75 countries. Here’s what it means for travelers

    January 16, 2026
  • Recipes

    winter cabbage salad with mandarins and cashews

    December 19, 2025

    pumpkin basque cheesecake

    November 25, 2025

    crunchy brown butter baked carrots

    November 19, 2025

    baked potatoes with crispy broccoli and bacon

    October 30, 2025

    brown butter snickerdoodles

    October 21, 2025
Gossips Today
  • Tech & Innovation
  • Healthcare
  • Personal Finance
  • Lifestyle
  • Travel
  • Business
  • Recipes
Business & Entrepreneurship

4 principles for using AI to spot abuse—without making it worse

gossipstodayBy gossipstodayJune 18, 2025No Comments6 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
4 principles for using ai to spot abuse—without making it worse
Share
Facebook Twitter LinkedIn Pinterest Email

Artificial intelligence is rapidly being adopted to help prevent abuse and protect vulnerable people—including children in foster care, adults in nursing homes, and students in schools. These tools promise to detect danger in real time and alert authorities before serious harm occurs.

Developers are using natural language processing, for example—a form of AI that interprets written or spoken language—to try to detect patterns of threats, manipulation, and control in text messages. This information could help detect domestic abuse and potentially assist courts or law enforcement in early intervention. Some child welfare agencies use predictive modeling, another common AI technique, to calculate which families or individuals are most “at risk” for abuse.

When thoughtfully implemented, AI tools have the potential to enhance safety and efficiency. For instance, predictive models have assisted social workers to prioritize high-risk cases and intervene earlier.

But as a social worker with 15 years of experience researching family violence—and five years on the front lines as a foster-care case manager, child abuse investigator, and early childhood coordinator—I’ve seen how well-intentioned systems often fail the very people they are meant to protect.

Now, I am helping to develop iCare, an AI-powered surveillance camera that analyzes limb movements—not faces or voices—to detect physical violence. I’m grappling with a critical question: Can AI truly help safeguard vulnerable people, or is it just automating the same systems that have long caused them harm?

New tech, old injustice

Many AI tools are trained to “learn” by analyzing historical data. But history is full of inequality, bias, and flawed assumptions. So are people, who design, test, and fund AI.

That means AI algorithms can wind up replicating systemic forms of discrimination, like racism or classism. A 2022 study in Allegheny County, Pennsylvania, found that a predictive risk model to score families’ risk levels—scores given to hotline staff to help them screen calls—would have flagged Black children for investigation 20% more often than white children, if used without human oversight. When social workers were included in decision-making, that disparity dropped to 9%.

Language-based AI can also reinforce bias. For instance, one study showed that natural language processing systems misclassified African American Vernacular English as “aggressive” at a significantly higher rate than Standard American English—up to 62% more often, in certain contexts.

Meanwhile, a 2023 study found that AI models often struggle with context clues, meaning sarcastic or joking messages can be misclassified as serious threats or signs of distress.

These flaws can replicate larger problems in protective systems. People of color have long been over-surveilled in child welfare systems—sometimes due to cultural misunderstandings, sometimes due to prejudice. Studies have shown that Black and Indigenous families face disproportionately higher rates of reporting, investigation, and family separation compared with white families, even after accounting for income and other socioeconomic factors.

Many of these disparities stem from structural racism embedded in decades of discriminatory policy decisions, as well as implicit biases and discretionary decision-making by overburdened caseworkers.

Surveillance over support

Even when AI systems do reduce harm toward vulnerable groups, they often do so at a disturbing cost.

In hospitals and eldercare facilities, for example, AI-enabled cameras have been used to detect physical aggression between staff, visitors, and residents. While commercial vendors promote these tools as safety innovations, their use raises serious ethical concerns about the balance between protection and privacy.

In a 2022 pilot program in Australia, AI camera systems deployed in two care homes generated more than 12,000 false alerts over 12 months—overwhelming staff and missing at least one real incident. The program’s accuracy did “not achieve a level that would be considered acceptable to staff and management,” according to the independent report.

Children are affected, too. In U.S. schools, AI surveillance like Gaggle, GoGuardian, and Securly are marketed as tools to keep students safe. Such programs can be installed on students’ devices to monitor online activity and flag anything concerning.

But they’ve also been shown to flag harmless behaviors—like writing short stories with mild violence, or researching topics related to mental health. As an Associated Press investigation revealed, these systems have also outed LGBTQ+ students to parents or school administrators by monitoring searches or conversations about gender and sexuality.

Other systems use classroom cameras and microphones to detect “aggression.” But they frequently misidentify normal behavior like laughing, coughing, or roughhousing—sometimes prompting intervention or discipline.

These are not isolated technical glitches; they reflect deep flaws in how AI is trained and deployed. AI systems learn from past data that has been selected and labeled by humans—data that often reflects social inequalities and biases. As sociologist Virginia Eubanks wrote in Automating Inequality, AI systems risk scaling up these long-standing harms.

Care, not punishment

I believe AI can still be a force for good, but only if its developers prioritize the dignity of the people these tools are meant to protect. I’ve developed a framework of four key principles for what I call “trauma-responsive AI.”

Survivor control: People should have a say in how, when, and if they’re monitored. Providing users with greater control over their data can enhance trust in AI systems and increase their engagement with support services, such as creating personalized plans to stay safe or access help.

Human oversight: Studies show that combining social workers’ expertise with AI support improves fairness and reduces child maltreatment—as in Allegheny County, where caseworkers used algorithmic risk scores as one factor, alongside their professional judgment, to decide which child abuse reports to investigate.

Bias auditing: Governments and developers are increasingly encouraged to test AI systems for racial and economic bias. Open-source tools like IBM’s AI Fairness 360, Google’s What-If Tool, and Fairlearn assist in detecting and reducing such biases in machine learning models.

Privacy by design: Technology should be built to protect people’s dignity. Open-source tools like Amnesia, Google’s differential privacy library, and Microsoft’s SmartNoise help anonymize sensitive data by removing or obscuring identifiable information. Additionally, AI-powered techniques, such as facial blurring, can anonymize people’s identities in video or photo data.

Honoring these principles means building systems that respond with care, not punishment.

Some promising models are already emerging. The Coalition Against Stalkerware and its partners advocate to include survivors in all stages of tech development—from needs assessments to user testing and ethical oversight.

Legislation is important, too. On May 5, 2025, for example, Montana’s governor signed a law restricting state and local government from using AI to make automated decisions about individuals without meaningful human oversight. It requires transparency about how AI is used in government systems and prohibits discriminatory profiling.

As I tell my students, innovative interventions should disrupt cycles of harm, not perpetuate them. AI will never replace the human capacity for context and compassion. But with the right values at the center, it might help us deliver more of it.

Aislinn Conrad is an associate professor of social work at the University of Iowa.

This article is republished from The Conversation under a Creative Commons license. Read the original article.

making principles spotabusewithout worse
Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticlePolice shut down Cluely’s party, the ‘cheat at everything’ startup
Next Article This Shoe Brand Is Known to Have a Waitlist of 100K+ Shoppers—and It Just Dropped the Perfect Summer Sneakers
admin
gossipstoday
  • Website

Related Posts

This common security measure is draining your workforce

January 18, 2026

You’re banned from blocking Trump’s face on your national park pass—but there’s a work-around

January 17, 2026

FDA commissioner’s drug review plan sparks alarm across the agency

January 17, 2026
Leave A Reply Cancel Reply

Demo
Trending Now

Saudi Arabia is already living the future of healthcare

Zero-Based Budgeting: What It Is and How to Use It

Week in Review:  Meta reveals its Oakley smart glasses

This Florida City Gets 361 Days of Sunshine Per Year — and It Has a Buzzy Food Scene and an Iconic Pink Hotel

Latest Posts

Why Silicon Valley is really talking about fleeing California (it’s not the 5%)

January 18, 2026

Kaiser affiliates to pay $556M to resolve Medicare Advantage fraud allegations

January 18, 2026

I'm a Flight Attendant, and This Carry-on From Amazon Is My Secret to Fitting a Month's Worth of Clothes in 1 Bag

January 18, 2026

Subscribe to News

Subscribe to our newsletter and stay updated with the latest news and exclusive offers.

Advertisement
Demo
Black And Beige Minimalist Elegant Cosmetics Logo (4) (1)
Facebook X (Twitter) Pinterest Vimeo WhatsApp TikTok Instagram

Categories

  • Tech & Innovation
  • Health & Wellness
  • Personal Finance
  • Lifestyle & Productivity

Company

  • About Us
  • Contact Us
  • Advertise With Us

Services

  • Privacy Policy
  • Terms & Conditions
  • Disclaimer

Subscribe to Updates

© 2026 Gossips Today. All Right Reserved.

Type above and press Enter to search. Press Esc to cancel.