Just a sample of the Echomail archive
Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.
|    CONSPRCY    |    How big is your tinfoil hat?    |    2,445 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 1,642 of 2,445    |
|    Mike Powell to All    |
|    U.S. blocking state-level    |
|    18 Aug 25 09:35:31    |
      TZUTC: -0500       MSGID: 1376.consprcy@1:2320/105 2d087593       PID: Synchronet 3.21a-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       TID: SBBSecho 3.28-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       BBSID: CAPCITY2       CHRS: ASCII 1       FORMAT: flowed       The U.S. is blocking state AI regulation. Here's what that means for every       business              Date:       Mon, 18 Aug 2025 14:01:19 +0000              Description:       Congress halts state AI regulation, pushing companies to self-govern amid       rapid enterprise adoption.              FULL STORY       ======================================================================              Congress didn't just reshape tax codes with the "One Big Beautiful" bill; it       also quietly reshaped the future of artificial intelligence . A lesser-known       provision of the sweeping legislation is now on its way to becoming law: a       10-year freeze on state-level AI regulation.               In other words, no individual state can pass rules that govern how businesses       develop or use AI systems. The message is clear for companies rushing to        embed AI in daily operations: govern yourselves or risk learning the hard way       why guardrails matter. AI tools are showing up in every workflow. with or       without oversight.              AI isn't a side project anymore. It's already embedded in cybersecurity       platforms, CRMs , internal chat tools, reporting dashboards and       customer-facing products. Even mid-size organizations are training AI models       on proprietary data to speed up everything from supplier selection to        contract analysis.               However, the adoption curve has outpaced internal checks. Many teams are       greenlighting tools without understanding how they were trained, what data       they retain or how outputs are validated. IT leaders often discover AI use       well after it's already operational. This kind of shadow Ai creates a major       risk surface.               And now, with state-level oversight blocked for a decade, there's no outside       pressure forcing organizations to establish policies or baseline rules. This       shift pushes businesses to take even more responsibility for what happens       inside their walls.              Without guardrails, AI can drift; fast               AI models aren't static. Once deployed, they learn from new data, interact       with systems and influence decision-making. That's powerful but also       unpredictable.               Left unchecked, an AI-driven forecasting tool might rely too heavily on       outdated patterns, causing overproduction or supply chain bottlenecks. A       chatbot designed to streamline customer service could unintentionally        generate biased or off-brand responses.               Meanwhile, generative models trained on sensitive business documents can       inadvertently expose proprietary information in future prompts. For example,        a study released in January 2025 found that nearly 1 in 10 prompts used by       business users when interacting with generative AI (GenAI) tools could       inadvertently disclose sensitive data.               These aren't abstract dangers; they've already appeared in public incidents.       But it's not just PR damage that's at stake. AI errors can affect revenue,       data security and even legal exposure. The absence of regulatory pressure       doesn't make these issues go away it makes them easier to miss until they're       too big to ignore.              The smart play is internal governance: before you need it               Organizations are eager to integrate GenAI, with many teams already using       these powerful tools in daily operations. This rapid adoption means that just       passively monitoring things isn't enough; a strong governance structure is       crucial, one that can adapt as AI becomes more central to the business.               Setting up an internal AI governance council, ideally with leaders from IT,       security, compliance and operations, offers that vital framework. This        council isn't there to stop innovation. Its job is to bring clarity. It       typically reviews AI tools before they're rolled out, sets clear usage       policies and works with teams so they fully understand the benefits and        limits of the AI they're using.               This approach reduces unauthorized tool usage, makes auditing more efficient       and helps leadership steer AI strategy with confidence. However, for       governance to be effective, it must be integrated into broader enterprise       systems, not siloed in spreadsheets or informal chats.              GRC platforms can anchor AI governance              Governance, risk and compliance (GRC) platforms already help businesses        manage third-party risk, policy enforcement, incident response and internal       audits. They're now emerging as critical infrastructure for AI governance as       well.               By centralizing policies, approvals and audit trails, GRC platforms help       organizations track where AI is being used, which data sources are feeding        it, and how outputs are monitored over time. They also create a transparent,       repeatable process for teams to propose, evaluate and deploy AI tools with       oversight so innovation doesn't become improvisation.              Don't count on vendors to handle it for you              Many tools advertise AI features with a sense of built-in safety, which       includes privacy settings, explainable models and compliance-ready        dashboards. But too often, the details are left up to the user.               If a vendor-trained model fails, your team will likely bear the operational       and reputational costs. Businesses can't afford to treat third-party AI as       "set and forget." Even licensed tools must be governed internally, especially       if they're learning from company data or making process-critical decisions.              The bottom line               With the U.S. blocking states from setting their own rules, many assumed       federal regulation would follow quickly. However, the reality is more       complicated. Draft legislation exists, but timelines are fuzzy, and political       support is mixed.               In the meantime, every organization using AI is effectively writing its own       rulebook. That's a challenge and an opportunity, especially for companies        that want to build trust, avoid missteps and confidently lead.               The organizations that define their governance now will have fewer fire        drills later. They'll also be better prepared for whatever federal rules       eventually arrive because their internal structure won't need a last-minute       overhaul.               Because whether or not rules are enforced externally, your business still       depends on getting AI right.                This article was produced as part of TechRadarPro's Expert Insights channel       where we feature the best and brightest minds in the technology industry       today. The views expressed here are those of the author and are not       necessarily those of TechRadarPro or Future plc. If you are interested in       contributing find out more here:       https://www.techradar.com/news/submit-your-story-to-techradar-pro              ======================================================================       Link to news story:       https://www.techradar.com/pro/the-u-s-is-blocking-state-ai-regulation-heres-wh       at-that-means-for-every-business              $$       --- SBBSecho 3.28-Linux        * Origin: capitolcityonline.net * Telnet/SSH:2022/HTTP (1:2320/105)       SEEN-BY: 105/81 106/201 128/187 129/14 305 153/7715 154/110 218/700       SEEN-BY: 226/30 227/114 229/110 111 114 206 300 307 317 400 426 428       SEEN-BY: 229/470 664 700 705 266/512 291/111 320/219 322/757 342/200       SEEN-BY: 396/45 460/58 712/848 902/26 2320/0 105 304 3634/12 5075/35       PATH: 2320/105 229/426           |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca