home bbs files messages ]

Just a sample of the Echomail archive

Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.

   CONSPRCY      How big is your tinfoil hat?      2,445 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 1,314 of 2,445   
   Mike Powell to All   
   ChatGPT can remember more   
   20 Apr 25 09:50:00   
   
   TZUTC: -0500   
   MSGID: 1050.consprcy@1:2320/105 2c6a4463   
   PID: Synchronet 3.20a-Linux master/acc19483f Apr 26 202 GCC 12.2.0   
   TID: SBBSecho 3.20-Linux master/acc19483f Apr 26 2024 23:04 GCC 12.2.0   
   BBSID: CAPCITY2   
   CHRS: ASCII 1   
   ChatGPT can remember more about you than ever before  should you be worried?   
      
   Date:   
   Sat, 19 Apr 2025 19:30:00 +0000   
      
   Description:   
   ChatGPTs new memory features promise smarter, more personal responses  but at   
   what cost to privacy, control, and connection?   
      
   FULL STORY   
   ======================================================================   
      
   ChatGPTs memory used to be simple. You told it what to remember, and it   
   listened.    
      
   Since 2024, ChatGPT has had a memory feature that lets users store helpful   
   context. From your tone of voice and writing style to your goals, interests,   
   and ongoing projects. You could go into settings to view, update, or delete   
   these memories. Occasionally, it would note something important on its own.   
   But largely, it remembered what you asked it to. Now, thats changing.    
      
   OpenAI, the company behind ChatGPT, is rolling out a major upgrade to its   
   memory . Beyond the handful of facts you manually saved, ChatGPT will now    
   draw from all of your past conversations to inform future responses by    
   itself.    
      
    According to OpenAI , memory now works in two ways: saved memories, added   
   directly by the user, and insights from chat history, which are the ones that   
   ChatGPT will gather automatically.    
      
   This feature, called long-term or persistent memory, is rolling out to    
   ChatGPT Plus and Pro users. However, at the time of writing, its not    
   available in the UK, EU, Iceland, Liechtenstein, Norway, or Switzerland due    
   to regional regulations.    
      
   The idea here is simple: the more ChatGPT remembers, the more helpful it   
   becomes. Its a big leap for personalization. But its also a good moment to   
   pause and ask what we might be giving up in return.   
      
   A memory that gets personal   
      
   Its easy to see the appeal here. A more personalized experience from ChatGPT   
   means you explain yourself less and get more relevant answers. Its helpful,   
   efficient, and familiar.    
      
   Personalization has always been about memory, says Rohan Sarin, Product   
   Manager at Speechmatics , an AI speech tech company. Knowing someone for   
   longer means you dont need to explain everything to them anymore.    
      
   He gives an example: ask ChatGPT to recommend a pizza place, and it might   
   gently steer you toward something more aligned with your fitness goals  a   
   subtle nudge based on what it knows about you. It's not just following   
   instructions, its reading between the lines.    
      
   Thats how we get close to someone, Sarin says. Its also how we trust them.   
   That emotional resonance is what makes these tools feel so useful  maybe even   
   comforting. But it also raises the risk of emotional dependence. Which,   
   arguably, is the whole point.    
      
   From a product perspective, storage has always been about stickiness, Sarin   
   tells me. It keeps users coming back. With each interaction, the switching   
   cost increases.    
      
   OpenAI doesnt hide this. The company's CEO,. Sam Altman, tweeted that memory   
   enables AI systems that get to know you over your life, and become extremely   
   useful and personalized.    
      
   That usefulness is clear. But so is the risk of depending on them not just to   
   help us, but to know us.   
      
   Does it remember like we do?   
      
   A challenge with long-term memory in AI is its inability to understand    
   context in the same way humans do.    
      
   We instinctively compartmentalize, separating whats private from whats   
   professional, whats important from whats fleeting. ChatGPT may struggle with   
   that sort of context switching.    
      
   Sarin points out that because people use ChatGPT for so many different    
   things, those lines may blur. IRL, we rely on non-verbal cues to prioritize.   
   AI doesnt have those. So memory without context could bring up uncomfortable   
   triggers.    
      
   He gives the example of ChatGPT referencing magic and fantasy in every story   
   or creative suggestion just because you mentioned liking Harry Potter once.   
   Will it draw from past memories even if they're no longer relevant? Our   
   ability to forget is part of how we grow, he says. If AI only reflects who we   
   were, it might limit who we become.    
      
   Without a way to rank, the model may surface things that feel random,   
   outdated, or even inappropriate for the moment.   
      
   Bringing AI memory into the workplace   
      
   Persistent memory could be hugely useful for work. Julian Wiffen, Chief of AI   
   and Data Science at Matillion , a data integration platform with AI built in,   
   sees strong use cases: It could improve continuity for long-term projects,   
   reduce repeated prompts, and offer a more tailored assistant experience," he   
   says.    
      
   But hes also wary. In practice, there are serious nuances that users, and   
   especially companies, need to consider. His biggest concerns here are    
   privacy, control, and data security.    
      
   I often experiment or think out loud in prompts. I wouldnt want that retained    
    or worse, surfaced again in another context, Wiffen says. He also flags    
   risks in technical environments, where fragments of code or sensitive data   
   might carry over between projects, raising IP or compliance concerns. These   
   issues are magnified in regulated industries or collaborative settings.   
      
   Whose memory is it anyway?   
      
   OpenAI stresses that users can still manage memory  delete individual    
   memories that aren't relevant anymore, turn it off entirely, or use the new   
   Temporary Chat button. This now appears at the top of the chat screen for   
   conversations that are not informed by past memories and won't be used to   
   build new ones either.    
      
   However, Wiffen says that might not be enough. What worries me is the lack of   
   fine-grained control and transparency, he says. It's often unclear what the   
   model remembers, how long it retains information, and whether it can be truly   
   forgotten.    
      
   Hes also concerned about compliance with data protection laws, like GDPR:    
   Even well-meaning memory features could accidentally retain sensitive    
   personal data or internal information from projects. And from a security   
   standpoint, persistent memory expands the attack surface. This is likely why   
   the new update hasn't rolled out globally yet.    
      
   Whats the answer? We need clearer guardrails, more transparent memory   
   indicators, and the ability to fully control whats remembered and whats not,"   
   Wiffen explains.   
      
   Not all AI remembers the same   
      
   Other AI tools are taking different approaches to memory. For example, AI   
   assistant Claude doesnt store persistent memory outside your current   
   conversation. That means fewer personalization features, but more control and   
   privacy.    
      
    Perplexity , an AI search engine, doesnt focus on memory at all  it    
   retrieves real-time web information instead. Whereas Replika, AI designed for   
   emotional companionship, goes the other way, storing long-term emotional   
   context to deepen relationships with users.    
      
   So, each system handles memory differently based on its goals. And the more   
   they know about us, the better they fulfill those goals  whether thats    
   helping us write, connect, search, or feel understood.    
      
   The question isnt whether memory is useful; I think it clearly is. The   
   question is whether we want AI to become this good at fulfilling these roles.    
      
   Its easy to say yes because these tools are designed to be helpful,    
   efficient, even indispensable. But that usefulness isnt neutral, its   
   intentional. These systems are built by companies that benefit when we rely    
   on them more.    
      
   You wouldnt willingly give up a second brain that remembers everything about   
   you, possibly better than you do. And thats the point. Thats what the   
   companies behind your favorite AI tools are counting on.   
      
   ======================================================================   
   Link to news story:   
   https://www.techradar.com/computing/artificial-intelligence/chatgpt-can-rememb   
   er-more-about-you-than-ever-before-should-you-be-worried   
      
   $$   
   --- SBBSecho 3.20-Linux   
    * Origin: capitolcityonline.net * Telnet/SSH:2022/HTTP (1:2320/105)   
   SEEN-BY: 105/81 106/201 128/187 129/305 153/7715 154/110 218/700 226/30   
   SEEN-BY: 227/114 229/110 111 114 206 300 307 317 400 426 428 470 664   
   SEEN-BY: 229/700 705 266/512 291/111 320/219 322/757 342/200 396/45   
   SEEN-BY: 460/58 712/848 902/26 2320/0 105 3634/12 5075/35   
   PATH: 2320/105 229/426   
      

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca