The Kids Online Safety Act (KOSA) would censor the internet and would make government officials the arbiters of what young people can see online. It will likely lead to age verification, handing more power, and private data, to third-party identity verification companies like Clear or ID.me. The government should not have the power to decide what topics are “safe” online for young people, and to force services to remove and block access to anything that might be considered unsafe for children. This isn’t safety—it’s censorship.

  • morrowind@lemmy.ml
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    I think we’ll just have to wait and see how tech companies implement this and how it’s enforced. Even the study is, as the letter points out, just guidance and not enforceable and can be ignored. The bill itself contains very little beyond saying that it doesn’t explicitly enforce “age gating” and extra data collection to determine age.

    Also, as the letter itself points out

    To date, COPPA has had negligible effects on adults because services directed to children under 13 are unlikely to be used by anyone other than children due to their limited functionality, effectively mandated by COPPA. But extending COPPA’s framework to sites “directed to” older teens would significantly burden the speech of adults because the social media services and games that older teens use are largely the same ones used by adults.

    Would it be impossible to create separation between sites used by older teens and adults? A lot of it happens culturally anyway. I’m not as pessimistic as others are about this.

    • MiscreantMouse@kbin.socialOP
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Would it be impossible to create separation between sites used by older teens and adults?

      Obviously it’s not impossible, it just requires sites to obtain a verifiable proof of age, i.e., a government ID.

      A lot of pathological optimism in this thread, and it might not impact you (at first), but the document you’re quoting explains why a lot of people are concerned:

      KOSA would require online services to “prevent” a set of harms to minors, which is effectively an instruction to employ broad content filtering to limit minors’ access to certain online content. Content filtering is notoriously imprecise; filtering used by schools and libraries in response to the Children’s Internet Protection Act has curtailed access to critical information such as sex education or resources for LGBTQ+ youth. Online services would face substantial pressure to over-moderate, including from state Attorneys General seeking to make political points about what kind of information is appropriate for young people. At a time when books with LGBTQ+ themes are being banned from school libraries and people providing healthcare to trans children are being falsely accused of “grooming,” KOSA would cut off another vital avenue of access to information for vulnerable youth.

      • morrowind@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Not disagreeing with you but where are you quoting that from? I don’t see it in the letter. Am I looking at the wrong thing?

        Also there is the provision to not censor anything that minors search for. I’m not saying the law is perfect mind you but

        • MiscreantMouse@kbin.socialOP
          link
          fedilink
          arrow-up
          2
          ·
          edit-2
          1 year ago

          Oops, you’re absolutely right about the attribution, the quote I posted above is from an earlier letter, I had too many open at once.

          Unfortunately, the provision you mention is essentially a bad-faith attempt to skirt the first amendment objections, while leveraging the imposed ‘duty of care’ to allow State AGs to censor with impunity. From p.6 of the more recent letter:

          KOSA will enable politically motivated actors to purge the Internet of speech that they dislike under the guise of “protecting minors.” Section 11(b) permits state attorneys general to bring enforcement actions whenever they believe that a resident of their state has been adversely affected by an alleged violation of KOSA. The inevitable abuse is entirely predictable. Consider two possibilities. First, in the aftermath of the May 14, 2022 mass shooting in Buffalo, New York Attorney General Letitia James issued a report blaming social media platforms for hosting the hateful speech that radicalized the shooter and calling for increased civil liability.27 Under KOSA’s duty of care, James could file suit 28 alleging failure to mitigate or prevent “physical violence” that might affect a minor user to pressure platforms into removing any speech deemed “hateful.”

          Second, some have already admitted that KOSA will be 29 used to censor LGBTQ content, especially that which relates to gender-affirming care. Armed with cherry-picked and selectively interpreted studies associating trans content with “anxiety, depression . . . and suicidal behavior,” 30 an ambitious attorney general will claim that “evidence-informed medical information” requires that platforms prohibit minors from viewing such content under KOSA’s duty of care. A state attorney general need not win a lawsuit—or even file one at all—to effectuate censorship. They need only initiate a burdensome investigation to pressure platforms to take down or restrict access to disfavored content.(31)