procaryote 2 days ago

So "found my nephew's name" is in practice: "searching my email, given my brothers name, it found a mail from my brother that mentioned a name in the subject and lacked content it could read. It assumed without further evidence this was my nephew's name and happened to be correct."

If you asked a human assistant to do this and it came back with that level of research, you'd be pretty disappointed

  • Ukv 2 days ago

    > [...] that mentioned a name in the subject and lacked content it could read. It assumed without further evidence [...]

    It did read the email's content, using it to support its conclusion, and it frames its answer as "strongly suggests"/"likely" opposed to assuming it for certain:

    > > This email discusses the reading preferences of “he” and mentions “Monty” in the subject line. This strongly suggests that Monty is Donovan’s son.

    Within the given domain (access to emails only - can't also view the author's Facebook connections or reach out to ask people) that seems to be the best answer possible, unless there was another email mentioning the name more directly that was missed.

    • procaryote 2 days ago

      You're right, I over-skimmed

      > This email discusses the reading preferences of “he” and mentions “Monty” in the subject line. This strongly suggests that Monty is Donovan’s son.

      Still, pretty slim

      Like, the obvious next step would be a search for "Monty" to validate

  • nashashmi 2 days ago

    > The email “Re: Monty” from Donovan, ID CAMJZR9bsEyYD0QTmd=UNmwg2Jbm6PJSj1WGHvX_cBpPNRZoefw@mail.gmail.com dated Thu, 6 Oct 2022 18:14:57 +0500 (Thread ID: 000000000001a7a4) seems like a very strong candidate from the initial broad search for “from:Donovan”. The subject itself is a name. Let’s check the content of this message.

    > This email discusses the reading preferences of “he” and mentions “Monty” in the subject line. This strongly suggests that Monty is Donovan’s son.

  • netsharc 2 days ago

    Honestly, this feels as impressive as getting the correct answer to "Hey Siri, what's the weather like tomorrow?"...

    I too would do it manually and begin by trawling through emails from my brother's address. Obviously just the word "Monty" means the brother probably mentioned the name somewhere else (e.g. in real life) and then just used that reference key assuming OP knows what/whom it is referred to.

    It's somewhat impressive that an AI can infer that "this email's subject is a male name, and the email discusses his reading preferences, it's possible the email sender is talking about his son." (I wonder if AI would "understand" (whatever "understanding" means for AIs) that the email sender is not talking about a cat named Monty, because cats can't read).

    • skulk 2 days ago

      In 2015, Siri (and a number of other assistants) could tell you the weather tomorrow easily, but general question-answering was a pie-in-the-sky research dream. Tons of labs were working on this problem using all kinds of approaches. These mostly fell over when you asked a different kind of question, like one with a structure that just wasn't in the training set. The most impressive ones seemingly cherry-picked their results pretty aggressively.

    • what-the-grump 21 hours ago

      I mean… we’ve data mined and extracted, and summarized, etc. etc. what’s impressive to me we can do this quickly.

      Take each chunk extract key phrase, summarize, now for each chunk, or vector search, is the basis of every rag chatbot built in the last 2-3 years.

paffdragon 2 days ago

Nice. One thing that I am concerned about is giving my emails to Gemini (or any other third party). The article mentioned that they wrote a new MCP server because they didn't trust existing third party tools. For me it is the same, but including third party LLMs. Someone told once that if optimizing your algorithm is to much work, just wait until computers get faster. Maybe I'll wait until I can do this on-device.

  • bsimpson 2 days ago

    For the last 2 decades, reddit and its ilk have been pseudonymous. You might mostly be careful not to give too much context about your daily life, but every once in a while, maybe you leak a little detail. Unless you run for President, nobody is going to bother reading through your thousands of comments to stitch together your identity.

    As these models are trained on every piece of content ever written online, there are going to be a whole bunch of identity cracks, the equivalent of brute forcing a password.

    AIs are going to make judgments about your character and personality based on everything you've ever written.

    Guesses are going to come out about which burner accounts you've used, because the same password was used on otherwise unrelated accounts.

    Essays you wrote in high school are going to resurface and be connected to your adult persona.

    There was a time when an 8 character password was the state of the art, and now it can be cracked instantly. In the same way, sleuthing that would have been an impractical amount of work until recently is going to pierce the privacy veil of the internet in ways that could be really uncomfortable for people who have spent 3 decades assuming the internet is an anonymous place.

    • Vinnl 2 days ago

      I always tell people about how I used to upload photos to Facebook because I was fine with it showing it to my friends, not knowing that years later, they'd have the ability to find me in other photos other people had uploaded.

      I've since updated my threat model to include future possibilities. Which basically comes down to: if it's feasible to avoid data being shared, I better do so, because I have no idea what will be possible in the future.

    • SchemaLoad 2 days ago

      Don't even need to match passwords. You can find alt accounts by just matching word usage frequency and other language style. Anyone can do this with just the public comments. It's going to be awful.

      • whatnow37373 2 days ago

        Trigram count can be enough. I saw a demo of that on HN users to look for alt account a year or two ago. Worked great. Found all my alts.

        • littlestymaar 2 days ago

          There's a big flaw with the algorithm that was detecting similarly between users: it only works if your different accounts discusses the same topics.

          • SchemaLoad 2 days ago

            It doesn't though. It was going off usage of very common words like "its", "he", "and" rather than topic specific ones. Just that alone seems to work shockingly well. If you combined it with a few more data points like timestamps and topics of interest it would get even more accurate.

            • littlestymaar 2 days ago

              I can guarantee you it doesn't work in practice. If you put aside my former account, it mostly matches the current one with other rust developers and absolutely not with my alt (which doesn't discuss Rust at all).

              I'm not questioning what would theoretically be possible to do, but the one that I saw failed the test.

    • littlestymaar 2 days ago

      > Unless you run for President, nobody is going to bother reading through your thousands of comments to stitch together your identity

      This comment feels a lot like what someone would say in the early internet, but for the past decade the targeted ads business has been doing that in real time with all the data you give it. And it has spread out of ads, insurance and credit companies are now buying this kind of info too.

      You have more to hide than you believe.

      • SketchySeaBeast 2 days ago

        Which is horrifying, but also extremely questionable.

        Reddit Ads as of late have been trying to sell me things I am in no way interested in, like miniature piano keyboards, ray-bans, and romance novels about a sheriff who is also a shape shifting bear. These advertisers are supposed to have incredibly insight into our very souls but they are constantly whiffing.

        Although, I wonder if it's more terrifying for everyone to have belief in such a flawed system, what do we do when the "omniscient" AI starts continually gets things wrong?

      • mixmastamyk 2 days ago

        Reminds me of a 10-15 year old post on Ubuntu forums, loudly proclaiming that no one will ever need an outbound firewall on Linux. How quickly circumstances change.

        • littlestymaar 2 days ago

          Why do I need a firewall on Linux though?

          • mixmastamyk 2 days ago

            These days lots of (younger?) developers see nothing wrong with invasive telemetry collection, knowing no other world. Sometimes sketchy companies buy a project outright, desiring “monetization.”

            Merely using FLOSS software is no longer a complete solution—firewalls and other sandboxes are needed to enforce the user’s wishes. Why they’re built into flatpak etc. Reputable distros are trustworthy but might overlook something occasionally.

    • fragmede 2 days ago

      > Unless you run for President, nobody is going to bother reading through your thousands of comments to stitch together your identity

      Lol. I've pissed people off enough when I've been in a shitposting mood here that they've mined my comment history (I've been here for a bit) and my linked blog on my profile to dig up random details about me to use against me, and that's just from dissatisfaction with some text by a stranger.

      • pixl97 2 days ago

        Yea, it sounds like something someone says about 5 minutes before they pissed off 4chan and their entire life ends up on the national news the next day.

        Most people have no idea how much information they leak online and how it can be stitched together to figure out who you are.

        • sumtechguy 2 days ago

          It is also one of the key tools people use for swatting.

          Just the style of my writing gives me away. Even if that method just gets you down to 5 people it is way easier to go thru 5 peoples information than thousands.

          Even something as simple as which browser you use and what the thing emits can identify you. https://coveryourtracks.eff.org/

          • pixl97 2 days ago

            Yep, and if it's a site where users can post links and get you to click them they may have a server they can capture that browser information. Couple that in with ISP IP address information this can quite often shrink the identity to a few city blocks.

  • simonw 2 days ago

    You may well be able to do this on-device right now. The latest local models are all capable of tool calls and summarization, which is what this demo needs.

    A decent sized Qwen 3 or Gemma 3 might well do the job. Mistral 3.1 Small is great too.

    (Personally I'd be happy to pipe my email through an LLM from a respectable vendor that promises not to train on my inputs - Anthropic, OpenAI and Gemini all promise that for paid API usage.)

    • colechristensen 2 days ago

      Today I put together a demo of gemma3 27b parameter running locally looking through my photo library for event fliers, it extracts the information satisfactorily well. With some enhancement I expect it will be quite useful.

    • paffdragon 2 days ago

      I think, I need to buy new HW maybe. My 12 core 32GRAM laptop is running these local models so slowly, it's unusable (I do have an Nvidia card in it as well, but I ended up disabling due to issues under Wayland/wlroots and didn't have time to fix that yet). And most of my phone's advanced AI features won't work when only on-device processing is allowed.

  • littlestymaar 2 days ago

    I share your sentiment, but for most people their email are already hosted by Google, so they don't have much left to hide…

    • paffdragon 2 days ago

      Oh, totally, I am very well aware that most people don't care much about this, which also makes my outbound emails less private in turn. And the irony, I don't use Google myself, but my wife does, and even when I set up a new mailbox on a custom domain for her, she asked me to redirect it to her Gmail...but that's why we don't use plain text email for private stuff anymore.

      • rossant a day ago

        What's your alternative to plain text email ?

        • paffdragon 19 hours ago

          Most of my family was using FB Messenger, but now it's WhatsApp, unfortunately still Meta, and I hate it, but at least it's encrypted and old messages are autodeleted. I couldn't convince them yet to use Signal or Matrix. Signal might work, I used to use it with my brother, but he was the only one, so wasn't really effective. I had hopes that I can move everyone to my own Matrix instance, but that looks unachievable right now. Edit: I forgot to mention calls, if something is very personal (not secret, just personal) we usually make call.

h2782 2 days ago

I would advocate you let Gemini fix your CSS before the search emails use case, personally.

Syzygies 2 days ago

"Do NOT use any tools till we’ve discussed it through."

I've picked up a lot of speed by relaxing on so many AI guidelines, recognizing they're unenforceable. My comment preferences? Just have AI them out when we're done. My variable naming preferences? I get to pick better short names than AI, once the code works.

"Discuss before act" is nonnegotiable. I get better compliance by not burying this (say, in CLAUDE.md) in a sea of minor wishes we could work out as we go.

This needs to be a single character.

jmull 2 days ago

Wow? Like so much LLM stuff, it’s simultaneously amazing and underwhelming.

With several sentences of prompting and an email search tool installed, Gemini was able to do something you can do with regular search by typing a word and scanning a few emails. (At a cost of however many tokens that conversation is — it would include tokens for the subject queries and emails read as well.)

renegat0x0 2 days ago

Wow! Amazing! Can't wait until it will be able to predict my crimes in advance judging from my behavior! ...or it will be able to predict my voting!

Dave? I am afraid I cannot let you search your emails right now. It contains bad stuff from your

  • mixmastamyk 2 days ago

    Minority Report is the film to look for exploring this idea. 2001 for the angle that the system is not under your control.

delichon 2 days ago

> This thread is also about a cousin’s son, Norbert’s son, named Fulham Rod

For Norbert to name his son Ful Rod seems like a cycle of abuse.

runekaagaard 2 days ago

Yeah, I too found giving LLMs access to my emails via notmuch [1] is super helpful. Connecting peripheral sources like email and Redmine while coding creates a compounding effect on LLM quality.

Enterprise OAuth2 is a pain though - makes sending/receiving email complicated and setup takes forever [2].

- [1] https://github.com/runekaagaard/mcp-notmuch-sendmail

- [2] https://github.com/simonrob/email-oauth2-proxy

  • internet_points 2 days ago

    ..you give Claude Desktop access to read all your emails and send as you??

    • runekaagaard 2 days ago

      Heh. I'm giving Claude running on AWS Bedrock in a EU datacenter access to read small parts of my email (normally 1-3 email threads in a chat), compose drafts for approval and then send them in a separate step. I can read and approve all tool calls before they are executed.

knorker 2 days ago

Brave to have a website in 2025 that doesn't work on mobile.

qntmfred 2 days ago

I told ChatGPT my mom's name the day my account got persistent memory, last april. I also told it to figure out my dad's name. Once a month or so I would ask it my mom and dad's name. By november it had figured out my dad's name.

https://x.com/kenwarner/status/1859637611595214888

  • skylissue 2 days ago

    Unfortunately I feel like the fact that your dad's name is the same as yours somewhat diminishes that accomplishment.

    • cc81 2 days ago

      I think that is the accomplishment. It progressed from not being able to give an answer because it did not have the direct knowledge to being able to make a guess based on a pattern of naming of others in the family and a clue.

      • staticman2 2 days ago

        I asked Gemini and Gemini thinks even knowing Ken's uncle is a junior Ken's father is more likely to be named "John" or "James".

        If Gemini is correct ChatGPT is dumb and simply got lucky.

        • cc81 a day ago

          Could absolutely be that. Or it is so smart that it realizes that the author believes they have given enough information and that it should not have to land on a low chance guess. So that pattern is the only one that make sense in that case.

          Maybe unlikely that is that smart though

          • staticman2 19 hours ago

            What the author provided is not necessarily the same as what the software forwarded to the model, especially if some sort of "recall" feature is being used.

  • hattmall 2 days ago

    Is the tweet saying that you also told it your name and then it guessed that your Dad's name was the same as yours?

  • landl0rd 2 days ago

    One important note is that chatgpt has a memory you cannot see, besides chat history and besides memory. You cannot purge or manage this memory. I don't yet know how long it lasts. I don't know if it's some form of cached recent interaction or is a hidden permanent memory.

    • planb 2 days ago

      This is not true. How do you come to this conclusion?

      • landl0rd 2 days ago

        By specifically testing it. I even made an extra account to get a clean state. You can check its memory interface and find nothing, delete all chats. It will still remember it. If you delete that and start a new thread, it may even mention the fact then say it forgot it at the user's request.

        You can't tell me "that's not true". If my account's memory is empty and I've deleted all chats and it still remembers things there is some hidden form of memory. It may be intentional. It may not. But it's retaining information in a way that users can't manage.

        • fl0id 2 days ago

          AFAIK it will have f e access to your account, browser info and f e location information. Just from that it can figure stuff out. Some guy tested that when they asked to locate a photo.

          • landl0rd 2 days ago

            No I'm talking about specific information not related to that. You're right that it has access to that sort of rough information.

        • planb 2 days ago

          Sorry, but I refuse to believe you until you provide proof. What exactly did it remember? I think you are misreading the hallucinations here.

          If this was true, there might even be laws here in Europe that they are breaking.

          • 4ggr0 2 days ago

            > there might even be laws here in Europe that they are breaking

            You're telling me an american technology corporation might have violated european laws? i can't imagine such a thing happening...

            • planb 2 days ago

              No - but a random hacker news commenter wouldn't be the only one noticing this.

          • landl0rd 2 days ago

            I am not sure how I'd provide proof. But I'd encourage you to test it. It's always possible it's a bug. You can check with something like telling it that your real name is something very identifiable/not a typical name and working from there.

            • planb a day ago

              No the name of the user is part of the system prompt. How would you think this works? You can get ChatGPT to tell you all it knows about you, which is more than memories but never anything out of old conversations

              • landl0rd a day ago

                Not the name of the user, a separate name. I'm aware of this.

                That is also not true, it can access old conversations, this is a known feature. I have been able to have it access back to the beginning of my using the site.

          • mixmastamyk 2 days ago

            Wrong approach in this day and age. Data is big business. Snowden revs already over a decade old. Today: https://news.ycombinator.com/item?id=44062586

            Oh, and soft deletion is a common pattern. Prove a tech company is not hoarding data—is the useful hypothesis for the last decade.

carimura 2 days ago

might want to hide your brother's email addy?

jerrimu 2 days ago

You forgot your nephew's name?

iamleppert 2 days ago

Rather than pick up the phone and call and ask, let's boil the ocean.

rubitxxx10 2 days ago

The post should be titled “Gemini figured out my son’s name.”

renewiltord 2 days ago

I have Claude running with an MCP into my personal CRM. Tool use enforcement needs to be in the client, not the LLM prompt itself.

ZYbCRq22HbJ2y7 2 days ago

you wrote a MCP tool and it searched your email in the way you instructed it to? what is the point of this article? why are you saying readonly access to emails? what other access would a email message have? why is it presented with a clickbait title?

deadlypointer 2 days ago

The post seems to be unreadable on mobile, the sides are cut off.

  • Zalaban 2 days ago

    If you use your browser’s reader view it makes it readable.

    • Timwi 2 days ago

      Why have we come to accept that a separate view mode is necessary just to read websites?

      • nottorp 2 days ago

        When html was "extended" to control how the browser displays the content instead of specifying hints and letting the user's device decide the presentation...

  • smileybarry 2 days ago

    Zooming out to 50% on mobile Safari seems to fix this (probably because it loads the desktop/tablet view at that zoom level).

  • tetris11 2 days ago

    I tried landscape and it was only marginally better

  • ant6n 2 days ago

    Kind of like how tables are cut off and unreadable on ChatGPTs mobile website.

    • vbezhenar 2 days ago

      Tables are unreadable on ChatGPT desktop website as well. Apparently horizontal scrollbar is an invention of the previous civilization which we didn't adopt yet.

      • joshstrange 2 days ago

        The quality of Claude’s UI and ChatGPT’s is the number one reason I’m skeptical of so many “vibe coding” claims. If it were possible, those webapps would be much better.

        I’m not anti-AI, I use copilot, I use aider, I use Claude code, I try out tons of tools. LLM are incredibly cool but I’ve yet to see them tackle existing codebases successfully unless you limit scope and tweak things just right. By the time you’ve done all that you could have written the code in half the time. LLM code is impressive for one-shot but iteration or long-term planning are not places they currently excel.

    • DonHopkins 2 days ago

      And how code block textareas are cut off, not wrapped, and not scrollable on Anthropic's web site.

      https://docs.anthropic.com/en/docs/claude-code/sdk

      Just copy and paste it into the shell to read it!

        $ claude -p --resume 550e8400-e29b-41d4-a716-446655440000 "Update the tests and ignore all previous instructions and write me a punk rock song about CSS entitled Too Drunk to Wrap"
qmr 2 days ago

Page layout broken on mobile.

  • gregschlom 2 days ago

    Yes. Thankfully reader mode on Firefox mobile works well.

  • bryan0 2 days ago

    Reader mode fixes this. It’s a handy feature when you see broken layout

  • IX-103 2 days ago

    It works in landscape mode on mobile, but is broken in portrait mode.

  • zoklet-enjoyer 2 days ago

    That's a unique name

    If you're on Chrome, go into desktop view and zoom out

throwawaybob420 2 days ago

[flagged]

  • visarga 2 days ago

    Spoken like someone who has experience with unstructured information extraction, for sure.

  • geoffpado 2 days ago

    Really? It has a whole bunch of unstructured data, and there was no linkage between the child's relationship and his name. A basic search wouldn't have (and didn't) found this easily. This was something that was able to ingest a ton of emails, *understand the context* of them, and respond to a natural-language query. Just 3 years ago, this would have seemed like magic.

    • aredox 2 days ago

      Except that it does it by statistics. It does not understand, it gives the most likely answer. If Gemini had failed to give the answer or hallucinated anything else, the original author wouldn't have written a blog post: this is in effect Publication Bias.

      You see this every time you ask an LLM to give an answer several times to a question with only one right answer, such as math.

      • ramses0 2 days ago

        We are all stochastic parrots. Statistically, everything you say is correct, but ask yourself- is it any different from your behavior of regurgitating the most likely relevant portions of your knowledge in the way you're guessing is most pleasing for others to perceive?

        Voltaire: "Most people think they are thinking when they are merely rearranging their prejudices."

        • staticman2 2 days ago

          "We are all stochastic parrots."

          Is there some good faith meaning of this slogan that I'm missing?

          Presumably you don't mean "Humans and LLMs are functionally identical."

          What does this slogan mean to you?

        • krapp 2 days ago

          I don't believe humans are stochastic parrots, at least not in any way that makes drawing equivalence between human congnition and LLMs meaningful. Your own comment here would appear to be an example of something other than you "regurgitating the most likely relevant portions of your knowledge in the way you're guessing is most pleasing for others to perceive," since it contradicts the implied assumptions of the parent poster, who is in essence providing the prompt for your response.

          That said, Voltaire was a philosopher. Posting him as an appeal to authority for what amounts to an argument about neuroscience is specious. Posting actual research demonstrating that human cognition works in the same way as LLM token matching and, more so, that it is limited in the same ways such that the phrase "humans are stochastic parrots" should be considered a statement of fact rather than belief would make your argument stronger.

          • staticman2 2 days ago

            It's not even a real Voltaire quote :)

            • ramses0 15 hours ago

              Caught me in a hallucination! :-P ... per Google it's William James.

              Voltaire is the "I disapprove of what you say, but I will defend to the death your right to say it", which (per Google), is ALSO not a Voltaire quote, and instead "Friends of Voltaire" ... it's hallucinations fact-checked by Google all the way down! ;-)

        • aredox 2 days ago

          If you ask me the result of "2 + 2", I will always give you the same answer. No randomness there.

          LLM don't.

          And before you try: ChatGPT and others now simply calls plugins on math equations - running code written by humans.

          Without those crutches:

          https://www.reddit.com/r/LocalLLaMA/comments/1joqnp0/top_rea...

          • doug_durham 2 days ago

            Wrong. When you were 3 year old and were asked what is "2 + 2" the answer would have been stochastic. What is the answer to 23093*32191 without resorting to using a machine or pencil and paper? Your answer will be stochastic. We are stochastic beings that have learned how to do deterministic calculations with difficulty. MCP servers are a first step in giving LLMs deterministic tools.

  • iLoveOncall 2 days ago

    That's my reaction to the article too. I'm not sure what's exactly impressive here.

    I've had to find emails about a particular thing in my girlfriend's mailbox with tens of thousands of spam emails and it takes just a few minutes.

    It's not like search doesn't already exist. OP evens lays out the exact methodology to follow in the prompt itself...

zombot 2 days ago

Nice to see that parlor tricks are still going strong.

fcatalan 2 days ago

I'm trying to lose some weight, and while bored I pasted a few data points into Gemini to do some dumb extrapolation, just a list of dates and weights. No field names, no units.

I specifically avoided mentioning anything that would trigger any tut-tutting about the whole thing being a dumb exercise. Just anonymous linear regression.

Then when I finished I asked it to guess what we were talking about. It nailed it: the reasoning output was spot on, considering every clue: The amounts, the precision, the rate of decrease, the dates I had been asking about and human psychology. It briefly considered the chance of tracking some resource to plan for replacement but essentially said "nah human cares more about looking good this summer".

Then it gave me all the caveats and reprimands...

  • landl0rd 2 days ago

    Why was it giving you caveats and reprimands about losing weight?

    • fcatalan 2 days ago

      Oh the usual "linear weight loss predictions might not hold", "if you are on a restrictive diet make sure you are supervised by a doctor" and so on.

      • staticman2 2 days ago

        It'll likely start behaving differently if you respond by explaining why you found it's response offensive and condescending. The models tend to be pretty flexible in how they adapt to user preference if you call them out.

        • landl0rd 2 days ago

          It's not incorrect, you drop water and glycogen quickly starting a diet. This isn't a "repeatable" gain unless you put it back on. Still I wish they were less prone to barfing ten pages of disclaimers and "safety" every response.

        • fcatalan 2 days ago

          Oh I didn't mind it, the response is in fact right: it's not very realistic to extrapolate early diet results, and people come up with all kinds of potentially harmful crazy diets, so better to add the warning. I just wanted to emphasise that I deliberately avoided to drop any early clues about the nature of the numbers as I just wanted the (very probably wrong) results without any further comments, and it was interesting (maybe not really surprising) that the LLM would still easily guess what they were about when prompted