home bbs files messages ]

Just a sample of the Echomail archive

Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.

   CONSPRCY      How big is your tinfoil hat?      2,445 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 2,439 of 2,445   
   Mike Powell to All   
   From Turing's ideas to Dartmouth researc   
   18 Feb 26 09:52:10   
   
   TZUTC: -0500   
   MSGID: 2197.consprcy@1:2320/105 2dfb10af   
   PID: Synchronet 3.21a-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0   
   TID: SBBSecho 3.28-Linux master/123f2d28a Jul 12 2025 GCC 12.2.0   
   BBSID: CAPCITY2   
   CHRS: ASCII 1   
   FORMAT: flowed   
   The $13,500 that changed the fate of humanity: how the term Artificial   
   Intelligence was first coined 71 years ago - but sadly without the legendary   
   visionary soul who imagined it   
      
   By Wayne Williams published 21 hours ago   
      
   From Turing's ideas to a Dartmouth research project, the origins of AI are   
   fascinating   
      
   Although AI may still feel like something new, the term itself was born more   
   than seven decades ago, during a modest proposal for a summer research project   
   at Dartmouth that carried a budget request of $13,500.   
      
   That proposal, submitted to the Rockefeller Foundation in 1955, marked the   
   first known appearance of the phrase "artificial intelligence."   
      
   It was an academic document, not a manifesto, but it quietly laid the   
   foundation for one of the most consequential technological movements in human   
   history.   
      
   The sad irony is that the field's most famous philosophical ancestor, Alan   
   Turing, was already gone by this point.   
      
   Turing had asked the defining question years earlier - "can machines   
   think?" - and designed what became known as the Turing Test, a method to   
   judge whether a machine could convincingly imitate human thought.   
      
   His work framed the entire discussion, yet he died in 1954, two years before   
   the Dartmouth meeting that officially named the field he had helped imagine.   
      
   Turing's death followed his prosecution in the UK for homosexuality, then   
   criminalized, and he died from cyanide poisoning in what was widely ruled a   
   suicide - a loss that removed one of computing's most original thinkers   
   just before his ideas began reshaping science.   
      
   Long before artificial intelligence had a name, Turing had already come up with   
   the question that would define it. In his 1950 paper Computing Machinery and   
   Intelligence, he proposed what became known as the Turing Test, or "imitation   
   game," replacing abstract debates about whether machines could truly think   
   with a simpler challenge: could a machine hold a written conversation well   
   enough that a human judge would be unable to reliably tell it apart from   
   another human?   
      
   By focusing on observable behavior instead of philosophy, Turing turned   
   intelligence into something researchers could actually test.   
      
   The idea was strikingly forward-looking given the reality of computers at the   
   time. Early machines were slow, expensive and limited to mathematical   
   calculation, yet Turing suspected that intelligence might emerge from   
   sufficiently complex symbol processing.   
      
   Rather than asking whether machines possessed a mind or consciousness, he asked   
   whether they could convincingly imitate intelligent behavior - something that   
   inspired later researchers to treat thinking as an engineering problem.   
      
   That conceptual leap directly influenced the group that gathered at Dartmouth   
   just a few years later, even though the man who posed the question would never   
   see the field formally named.   
      
   The Dartmouth Summer Research Project on Artificial Intelligence, organized by   
   John McCarthy with Marvin Minsky, Claude Shannon, and Nathaniel Rochester, was   
   small and ambitious.   
      
   According to the proposal, researchers hoped to prove that "every aspect of   
   learning or any other feature of intelligence can in principle be so precisely   
   described that a machine can be made to simulate it." The goal sounded   
   ambitious then and still does now: language, abstraction, reasoning, and   
   self-improvement, all encoded into machines.   
      
   McCarthy would later become one of AI's most influential voices. In a 1979   
   issue of ComputerWorld, he said bluntly that the computer revolution   
   "hasn't happened yet," even while predicting that it eventually would.   
      
   He argued that computers had not yet impacted life in the way electricity or   
   automobiles had, but he believed that applications in the coming decade would   
   initiate a genuine revolution.   
      
   McCarthy's realism often contrasted with the hype that surrounded the field,   
   a tension that has followed AI ever since.   
      
   Alan Turing: The Scientist Who Saved The Allies - https://youtu.be/XGqbieVcjPU   
      
   AI as a hot topic   
      
   By the early 1980s, interest in AI had surged again, but confusion about what   
   it really meant was widespread.   
      
   Writing in a 1984 issue of InfoWorld, reporter Peggy Watt noted that artificial   
   intelligence had become a "hot topic," with shelves filled with books and   
   software companies racing to label products as intelligent. Yet she warned that   
   "the term is being used and abused widely, almost to the point of losing its   
   usefulness as a description."   
      
   The frustration among researchers was obvious. In that same InfoWorld report,   
   Dr. S. Jerrold Kaplan of Teknowledge said, "Whenever anybody says, `I'm   
   selling AI,' I'm suspicious."   
      
   Kaplan argued that AI was not a single program. "The science of AI is a set   
   of techniques for programming," he said, describing systems that represented   
   "concepts and ideas, explanations and relationships," rather than just   
   numbers or words.   
      
   This tension between promise and reality also defined the work of Marvin   
   Minsky, one of Dartmouth's original architects. In a 1981 issue of   
   ComputerWorld, covering the Data Training '81 conference, Minsky described AI   
   as fundamentally paradoxical: "Hard things are easy to do and easy things are   
   hard to do."   
      
   Computers excelled at calculations that challenged humans, but struggled with   
   common sense, language ambiguity, and contextual understanding.   
      
   Minsky explained that "common sense is the most difficult thing to inculcate   
   into a computer."   
      
   Humans absorb countless exceptions and nuances over years of living, but   
   machines require explicit instruction. A logical rule like "birds can fly"   
   breaks down immediately when confronted with dead birds or flightless species   
   - a simple example revealing why intelligence is more than pure logic.   
      
   Expert systems   
      
   The optimistic early years of AI had already produced striking milestones. The   
   Lawrence Livermore National Laboratory later described how researchers in the   
   1960s developed programs such as SAINT, an early "expert system" capable of   
   solving symbolic integration problems at the level of a college freshman.   
      
   The program solved nearly all the test problems it faced, hinting that machines   
   could emulate specialist reasoning long before modern machine learning.   
      
   Yet progress came in waves. Funding boomed in the 1960s as government agencies   
   backed ambitious research, then cooled massively in the 1970s.   
      
   The dream of building human-like intelligence proved far harder than expected.   
   Even McCarthy admitted that "human-level" AI was still "several   
   conceptual revolutions away."   
      
   By the time AI returned to the spotlight in the 1980s, companies were marketing   
   expert systems and natural-language tools as breakthroughs.   
      
   Some systems impressed users by tolerating spelling mistakes or translating   
   plain English commands into database queries.   
      
   Others, however, leaned more on clever engineering than genuine reasoning. As   
   one unnamed researcher quoted in InfoWorld warned, the real test of an expert   
   system was whether it could explain its conclusions.   
      
   Still, the vision persisted. Industry observers imagined computers capable of   
   understanding natural language, translating documents, and even correcting   
   grammar automatically.   
      
   Kaplan predicted AI would change how people programmed because it was "much   
   more natural to work with symbolic terms than math algorithms." The idea that   
   software could assist, advise, and collaborate with humans was already taking   
   shape.   
      
   Looking back, what stands out is how many early predictions were both wrong and   
   right. McCarthy thought the revolution had not yet arrived, but he believed it   
   would come through practical applications. Minsky warned that common sense   
   would remain stubbornly difficult.   
   Hmm   
      
   Today, as AI systems write text, generate images, and assist scientific   
   discovery, the echoes of those early conversations remain.   
      
   The Dartmouth organizers imagined machines that could "use language, form   
   abstractions and concepts, solve kinds of problems now reserved for humans, and   
   improve themselves." All of which are (mostly) true today.   
      
   The $13,500 proposal did not seem remarkable at the time. It was just one   
   funding request among many. Yet it gave a name to an idea that continues to   
   change society, shaped by optimism, frustration, paradox, and unresolved   
   questions.   
      
   And perhaps that is the real legacy of artificial intelligence. It began not as   
   a single invention, like the transistor or the microprocessor, but as a wager   
   that intelligence itself could be understood, described, and eventually   
   reproduced.   
      
   Seventy-one years later, humanity is still testing that idea, still arguing   
   about definitions, and still pursuing the vision imagined by twentieth-century   
   minds who believed thinking machines might one day become real.   
      
      
   https://www.techradar.com/pro/the-usd13-500-that-changed-the-fate-of-humanity-h   
   ow-the-term-artificial-intelligence-was-first-coined-71-years-ago-but-sadly-wit   
   hout-the-legendary-visionary-soul-who-imagined-it   
      
   $$   
   --- SBBSecho 3.28-Linux   
    * Origin: Capitol City Online (1:2320/105)   
   SEEN-BY: 105/81 106/201 128/187 129/14 305 153/7715 154/110 218/700   
   SEEN-BY: 226/30 227/114 229/110 134 206 300 307 317 400 426 428 470   
   SEEN-BY: 229/664 700 705 266/512 291/111 320/219 322/757 342/200 396/45   
   SEEN-BY: 460/58 633/280 712/848 902/26 2320/0 105 304 3634/12 5075/35   
   PATH: 2320/105 229/426   
      

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca