The Musings of Jaime David
The Musings of Jaime David
@jaimedavid.blog@jaimedavid.blog

The writings of some random dude on the internet

1,089 posts
1 follower

Tag: family friendly internet

  • Age by Algorithm: Why YouTube’s New AI Age Checks Raise Big Questions for Creators and Viewers Alike

    Age by Algorithm: Why YouTube’s New AI Age Checks Raise Big Questions for Creators and Viewers Alike

    As creators, we know that the digital landscape is constantly evolving — new tools, new guidelines, and yes, new rules about who can see what and when. YouTube’s latest move? Using artificial intelligence to guess a viewer’s age, not based on their birthday, but on their behavior.

    That’s right. YouTube recently announced that it’s rolling out an AI-powered age detection system in the U.S. This system will estimate whether a user is over or under 18 by looking at what they watch, what they search for, and how long they’ve had their account — regardless of the birthdate they entered.

    For creators, this raises a lot of questions.

    1. Will our videos reach the intended audience?
    If someone is misclassified as a minor, they might be automatically excluded from seeing our content — even if it’s not inappropriate. That means creators could lose out on engagement, visibility, and potential revenue due to something as abstract as an algorithmic guess.

    2. What happens if the system gets it wrong?
    The burden falls on users to prove their age with a credit card, government ID, or selfie. This isn’t just a hassle — it’s a potential privacy concern, especially for users who don’t feel comfortable sharing such personal data online.

    3. What about nuance?
    Not all content is clearly “for kids” or “for adults.” Sometimes, it’s educational. Sometimes, it’s artistic. Will AI understand the difference? Or will creators start censoring themselves to avoid being caught in the system’s net?

    This rollout comes on the heels of broader regulatory trends — like the Kids Online Safety Act (KOSA) and the UK’s Online Safety Act — which aim to protect minors online. And while those goals are important, creators and digital users alike are increasingly worried that the methods used to “protect” may lead to overreach, mistrust, or unintended harm.

    YouTube says this approach has worked well in other countries and will be tested with a small group of U.S. users first. But even so, it’s important for us — as creators, viewers, and digital citizens — to pay attention. AI isn’t perfect. And when it’s used to gatekeep access, influence algorithms, or reshape who sees our work, the stakes are higher than ever.

    Let’s keep the conversation going. Let’s stay informed. And most of all, let’s advocate for smart solutions that protect young users without punishing creativity, curiosity, or community.

  • Creators and Congress: Why I’m Keeping an Eye on New Changes to Internet Laws

    Creators and Congress: Why I’m Keeping an Eye on New Changes to Internet Laws

    As someone who creates content online, I’m always paying attention to how the internet is changing — not just in terms of trends or technology, but also in terms of laws and policies. Recently, there’s been a lot of buzz about something called the Congressional Creators Caucus, and it got me thinking about what this might mean for people like me — and for the people who watch, read, or listen to our work.

    The Congressional Creators Caucus was launched earlier this year, and it’s meant to give digital content creators a stronger voice in Washington, D.C. It was supported by MatPat — a name many YouTube fans will recognize from Game Theory — and his wife Stephanie. They’ve been involved in the world of online content for a long time, so in some ways, it makes sense that they’d want to help creators be heard at the policy level.

    The idea of Congress listening to creators might sound exciting. And in some ways, it is. Creators work incredibly hard — often for long hours, with little financial certainty — and we face real challenges with algorithms, content rules, monetization changes, and staying safe online. Having lawmakers recognize those challenges is a step in the right direction.

    But I’m also cautious. Alongside this new caucus, there’s a federal bill called the Kids Online Safety Act (KOSA) that’s getting attention. On the surface, KOSA is about making the internet safer for kids, which is something we can all agree is important. But like a lot of things in government, the way a bill is written matters just as much as what it’s trying to do.

    Some creators, advocates, and privacy experts are worried that KOSA could go too far. Depending on how the rules are enforced, it could lead to too much content being taken down, especially posts that talk honestly about mental health, identity, or growing up. Others are concerned that it could require websites and apps to collect more personal information to “verify age,” which raises questions about online privacy — something that matters to everyone, no matter your age.

    I don’t want to sound alarmist. These conversations are still happening, and nothing is set in stone. But I do think it’s fair for creators to ask questions and stay informed. If policies like these change how we’re allowed to post, what we can share, or how audiences can find us, it’s going to affect not just creators — but also the communities we’ve built with our audiences over time.

    This isn’t about being for or against something politically. It’s about making sure we don’t rush into decisions that could unintentionally hurt the very people we’re trying to protect. We need laws that make the internet safer without silencing important voices or putting up walls between creators and their supporters.

    As someone who cares deeply about creativity, connection, and communication, I’m hopeful we can find the right balance. But until then, I’ll keep watching and speaking up — because the internet has given creators a place to thrive, and we shouldn’t lose that.