Just a sample of the Echomail archive
Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.
|    CONSPRCY    |    How big is your tinfoil hat?    |    2,445 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 2,134 of 2,445    |
|    Mike Powell to All    |
|    Keep kids safe in AI 2/2    |
|    30 Dec 25 09:42:02    |
      TZUTC: -0500       MSGID: 1891.consprcy@1:2320/105 2db9221d       PID: Synchronet 3.21a-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       TID: SBBSecho 3.28-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       BBSID: CAPCITY2       CHRS: ASCII 1       FORMAT: flowed        (continued)              What needs to happen?              Ideally, protecting children would involve parents, schools, governments, and       tech companies all working together. But after years of slow progress on       social media regulation, its not hard to see why confidence in that happening       any time soon is low.               Many of the biggest problems could be addressed if the companies behind AI       tools and social platforms took more responsibility and enforced meaningful       safeguards. Tech companies need to be subject to urgent, meaningful        regulation if were going to protect children, Steele says. At the moment, far       too much responsibility is falling on families, schools, and the goodwill of       industry, and that simply isnt safe.               Bartuski agrees that companies should be doing far more. They have the money,       resources, and visibility to be able to do a lot more. Many social media       companies have used Foggs Persuasive Design to get kids habituated to be       lifelong users of their platforms. Tech companies do this on purpose, she       explains.               But this is where the tension lies. We can say tech companies should do more,       yet as the risks become clearer, corporate incentives are often moving in the       opposite direction. With the guardrails being removed from AI development       (specifically in the US), there are some (not all) companies that are using       that to their advantage, Bartuski says. She has already seen companies push       ahead with features they know are dangerous.               Even so, experts agree that certain steps would have an immediate and       significant impact. There need to be clear rules on what AI systems must not       be allowed to do, including creating sexualized images of children, promoting       self-harm, or using design features that foster emotional dependency, Steele       says.               This forms the basis of the Safe AI for Children Alliances Non-Negotiables       Campaign , which outlines three protections every child should have.        Alongside banning the creation of sexualized images of children, the campaign       states that AI must never be designed to make children emotionally dependent       and AI must never encourage children to harm themselves.               But relying on tech companies alone wont cut it. Independent oversight is       essential. This is why Briercliffe believes stronger external checks are       needed across the industry. There must be mandatory, independent, third-party       testing and evaluation before deployment, he says. We also need independent       oversight, transparency about how systems behave in real-world conditions,        and real consequences when companies fail to protect children.               And ultimately, this goes beyond individual platforms. This is ultimately a       question of societal responsibility, Tara says. We must set strong,       enforceable standards that ensure childrens safety comes before commercial       incentives.              What can parents do?               Even with regulations slow to catch up, parents shouldnt feel at a loss.        There are meaningful steps you can take right now. Its completely       understandable for parents to feel worried, Steele says. The technology is       moving very fast, and the risks arent intuitive. But it is important not to       feel powerless.                1. Understand the basics               Parents dont need to learn how every AI tool works, Bartuski says. But        getting clear on the risks and benefits is important. Steele offers a free       Parent and Educator Guide at safeaiforchildren.org that lays out all the        major concerns in clear, accessible language, which is a good place to start.                2. Create open, non-judgmental communication               If kids feel judged or are worried about consequences, they are not going to       turn to parents when something is wrong, Bartuski says. If they dont feel        safe talking to you, you are placing them in potentially dangerous and/or       exploitative situations. Keep conversations calm, curious, and shame-free.                3. Talk about the tech               You might assume your children understand AI better than you do because they       use it more. But they may not grasp how it works, how often it gets things       wrong, or that fake content can look real. Bartuski says kids need to know       that chatbots can be wrong, manipulative, or unsafe, even when they sound       caring or convincing.                4. Use shared spaces               This isnt about banning tech outright. Its about making it safer. Steele       suggests enforcing shared spaces", which involves using AI tools in communal       areas, experimenting together, and avoiding private one-on-one use behind       closed doors. This could reduce the chance of harmful interactions going       unnoticed.                5. Extend the conversation beyond the home               Safety shouldnt stop at your front door. If you are worried, ask your child's       school what they have in place, Briercliffe says. Even ask your employer to       bring in a professional to give a talk. Experts agreed that while parents        play a key role here, this is a wider cultural challenge, and the more openly       we all discuss it, the safer children will be.                6. Find more balance and reduce screen time               Weve been talking about limiting screen time for years, and its just as       important now that AI is showing up across apps, games, and social platforms.       Kids need to be taught balance, Bartuski says. Play is essential for growth       and development. She also stresses that reducing screen time only works if        its replaced with activities that are engaging, fun, and cognitively       challenging.               ======================================================================       Link to news story:       https://www.techradar.com/ai-platforms-assistants/how-you-can-keep-your-kids-s       afe-in-this-ai-powered-world              $$       --- SBBSecho 3.28-Linux        * Origin: Capitol City Online (1:2320/105)       SEEN-BY: 105/81 106/201 128/187 129/14 305 153/7715 154/110 218/700       SEEN-BY: 226/30 227/114 229/110 134 206 300 307 317 400 426 428 470       SEEN-BY: 229/664 700 705 266/512 291/111 320/219 322/757 342/200 396/45       SEEN-BY: 460/58 633/280 712/848 902/26 2320/0 105 304 3634/12 5075/35       PATH: 2320/105 229/426           |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca