Just a sample of the Echomail archive
Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.
|    CONSPRCY    |    How big is your tinfoil hat?    |    2,445 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 2,426 of 2,445    |
|    Mike Powell to All    |
|    5 reasons why AI apocalypse might be clo    |
|    17 Feb 26 11:22:45    |
      TZUTC: -0500       MSGID: 2184.consprcy@1:2320/105 2df9d45d       PID: Synchronet 3.21a-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       TID: SBBSecho 3.28-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0       BBSID: CAPCITY2       CHRS: ASCII 1       FORMAT: flowed       'The world is in peril' - 5 reasons why the AI apocalypse might be closer       than you think              By Eric Hal Schwartz published 23 hours ago              AI is causing problems and there are warning signs it's going to get worse              There's been an endless parade of proclamations over the last few years about       an AI golden age. Developers proclaim a new industrial revolution and       executives promise frictionless productivity and amazing breakthroughs       accelerated by machine intelligence. Every new product seems to boast of its AI       capability, no matter how unnecessary.              But that golden sheen has a darker edge. There are more indications that the       issues around AI technology are not a small matter to be fixed in the next       update, but a persistent, unavoidable element of the technology and its       deployment. Some of the concerns are born out of myths about AI, but that       doesn't mean that there's nothing to worry about. Even if the technology isn't       scary, how people use it can be plenty frightening. And solutions offered by       the biggest proponents of AI solutions often seem likely to make things worse.              There have been events in the past few months that have hinted at something       more destabilizing. None of them guarantees catastrophe on their own, but they       don't evoke the optimism the fans of AI would like us to feel. They sketch a       picture of a technology accelerating faster than the structures meant to guide       it. If the apocalypse ever comes courtesy of artificial intelligence, they may       be what we look back at as the first moments.              1. AI safety experts flee              This month, the head of AI safety research at Anthropic resigned and did so       loudly. In a public statement, he warned that "the world is in peril" and       questioned whether core values were still steering the company's decisions. A       senior figure whose job was to think about the long term and how increasingly       capable systems might go wrong decided it was impossible to keep going. His       departure followed a string of other exits across the industry, including       founders and senior staff at xAI and other high-profile labs. The pattern has       been difficult to ignore.              Resignations happen in tech all the time, of course, but these departures have       come wrapped in moral concern. They have been accompanied by essays and       interviews that describe internal debates about safety standards, competitive       pressure, and whether the race to build more powerful models is outpacing the       ability to control them. When the people tasked with installing the brakes       begin stepping away from the vehicle, it suggests that the car may be       accelerating in ways even insiders find troubling.              AI companies are building systems that will shape economies, education, media,       and possibly warfare. If their own safety leaders feel compelled to warn that       the world is veering into dangerous territory, that warning deserves more than       a shrug.              2. Deepfake dangers              It's hard to argue there's an issue in AI safety when regulators in the United       Kingdom and elsewhere find credible evidence of horrific misuse of AI like       reports that Grok on X had generated sexually explicit and abusive imagery,       including deepfake content involving minors.              Not that deepfakes are new, but now, with the right prompts and a few minutes       of patience, users can produce fabricated images that would have required       significant technical expertise just a few years ago. Victims have little       recourse once manipulated images spread across the internet's memory.              From an apocalyptic perspective, this is not about a single scandal. It is       about erosion. Trust in visual evidence was already fragile. Now it is       cracking. If any image can be plausibly dismissed as synthetic and any person       can be digitally placed into compromising situations, the shared factual ground       beneath public discourse begins to dissolve. The hopeful counterpoint is that       regulators are paying attention and that platform operators are being forced to       confront misuse. Stronger safeguards, better detection tools, and clearer legal       standards could blunt the worst outcomes, but that won't undo the damage       already done.              3. Real world hallucinations              For years, most AI failures lived on screens, but that boundary is fading as AI       systems are starting to help steer cars, coordinate warehouse robots, and guide       drones. They interpret sensor data and make split-second decisions in the       physical world. And security researchers have been warning that these systems       are surprisingly easy to trick.              Studies have demonstrated that subtle environmental changes, altered road       signs, strategically placed stickers, or misleading text can cause AI vision       systems to misclassify objects. In a lab, that might mean a stop sign       interpreted as a speed limit sign. On a busy road, it could mean something far       worse. Malicious actors could be an emerging threat. The more infrastructure       depends on AI decision-making, the more attractive it becomes as a target.              The apocalyptic imagination leaps quickly from there. Ideally, awareness of       these weaknesses will drive investment in robust security practices before       autonomous systems become ubiquitous. But, if deployment outruns safety, we       could all learn some painful lessons in real time              4. Chatbots get a sales team              OpenAI began rolling out advertising within ChatGPT recently, and the response       has been mixed, to say the least. A senior OpenAI researcher resigning publicly       and arguing that ad-driven AI products risk drifting toward user manipulation       probably didn't help calm matters. When a system that knows your fears,       ambitions, and habits also carries commercial incentives, the lines between       assistance and persuasion blur.              Social media platforms once promised simple connections and information       sharing, but their ad-based business models shaped design choices that       maximized engagement, sometimes at the cost of user well-being. An AI assistant       embedded with advertising could face similar pressures. Subtle prioritization       of certain answers. Nudges that align with sponsor interests. Recommendations       calibrated not only for relevance but for revenue.              To be fair, advertising does not automatically corrupt a product. Clear       labeling, strict firewalls between monetization and core model behavior, and       regulatory oversight could prevent worst-case scenarios. Still, that       resignation means there's real, unresolved tension in that regard.              5. A growing catalogue of mishaps              There is a simple accumulation of evidence of problems with AI. Between       November 2025 and January 2026, the AI Incident Database logged 108 new       incidents. Each entry documents a failure, misuse, or unintended consequence       tied to AI systems. Many cases may be minor, but every report of AI used fraud       or dispensing dangerous advice adds up.              Acceleration is what matters here. AI tools are popping up everywhere, and so       the number of problems multiplies. Perhaps the uptick in reported incidents is       simply a sign of better tracking rather than worsening performance, but it's       doubtful that accounts for everything. A still relatively new technology has a       lot of harm to its name.              Apocalypse is a loaded word evoking images of collapse and finality. AI may not       be at that point and might never reach it, but it's undeniably causing a lot of       turbulence. Catastrophe is not inevitable, but complacency would be a mistake.       It might not be the end of the world, but AI can certainly make it feel that       way.                     https://www.techradar.com/ai-platforms-assistants/the-world-is-in-peril-5-reaso       ns-why-the-ai-apocalypse-might-be-closer-than-you-think              $$       --- SBBSecho 3.28-Linux        * Origin: Capitol City Online (1:2320/105)       SEEN-BY: 105/81 106/201 128/187 129/14 305 153/7715 154/110 218/700       SEEN-BY: 226/30 227/114 229/110 134 206 300 307 317 400 426 428 470       SEEN-BY: 229/664 700 705 266/512 291/111 320/219 322/757 342/200 396/45       SEEN-BY: 460/58 633/280 712/848 902/26 2320/0 105 304 3634/12 5075/35       PATH: 2320/105 229/426           |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca