• paddirn@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    1
    ·
    6 months ago

    It seems like such a weird thing to marry up with internet searching. This method where the algorithms can & will “hallucinate” and just make shit up vs finding very specific information that a person is searching for. Why ever trust these LLMs with facts? These things should’ve only ever been marketed for creative writing and art, not shit like writing legal briefs and school papers and such.

    • Blóðbók@slrpnk.net
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      2
      ·
      edit-2
      6 months ago

      Maybe I can share some insight into why one might want to.

      I hate searching the internet. It’s a massive mental drain for me to try figure out how I should put my problem into words that others with similar ideas will have done before me - it’s my mental processing power wasted on purely linguistic overhead instead of trying to understand and learn about the problem.

      I hate the (dis-/mis-)informational assault I open myself to by skimming through the results, because the majority of them will be so laughably irrelevant, if not actively malicious, that I become a slightly worse person every time I expose myself.

      And I hate visiting websites. Not only because of all the reasons modern websites suck, but because even if they are a delight in UX, they are distracting me from what I really want, which is (most of the time) information, not to experience someone’s idiosyncratic, artistic ideas for how to organise and present data, or how to keep me ‘engaged’.

      So yes, I prefer stupid a language model that will lie about facts half the time and bastardise half my prompts if it means I can glance a bit of what the internet has to say about something, because I can more easily spot plausible bullshit and discard it or quickly check its veracity than I can magic my vague problem into a suitable query only to sift through more ignorance, hostility, and implausible bullshit conjured by internet randos instead.

      And yes, LLMs really do suck even in their domain of speciality (language - because language serves a purpose, and they do not understand it), and they are all kinds of harmful, dangerous, and misused. Given how genuinely ignorant people are of what an LLM really is and what it is really doing, I think it’s irresponsible to embed one the way google has.

      I think it’s probably best to… uhh… sort of gatekeep this tech so that it’s mostly utilised by people who understand the risks. But capitalism is incompatible with niches and bespoke products, so every piece of tech has to be made with absolutely everyone as a target audience.

      • CrayonMaster@midwest.social
        link
        fedilink
        English
        arrow-up
        2
        ·
        6 months ago

        What are you searching for? I can’t remember the last time I googled something and most the results were malicious.

        Also, I don’t think it’ll be easier to spot bullshit coming from an LLM then a website.

        • Blóðbók@slrpnk.net
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          edit-2
          6 months ago

          I don’t know about google because I don’t use it unless I really can’t find what I’m looking for, but here’s a quick ddg search with a very unambiguous and specific question, and from sampling only the top 9 results I see 2 that are at all relevant (2nd and 5th):

          In order to answer my question, I need to first mentally filter out 7/9 of the results visible on my screen, then open both of the relevant ones in new tabs and read through lengthy discussions in order to find out if anyone has shared a proper solution.

          Here is the same search using perplexity’s default model (not pro, which is a lot better at breaking down queries and including relevant references):

          and I don’t have to verify all the details because even if some of it is wrong, it is immediately more useful information to me.

          I want to re-emphasise though that using LLMs for this can be incredibly frustrating too, because they will often insist assertively on falsehoods and generally act really dumb, so I’m not saying there aren’t pros and cons. Sometimes a simple keyword-based search and manual curation of the results is preferred to the nonsense produced by a stupid language model.

          Edit: I didn’t answer your question about malicious, but I can give some example of what I consider malicious and you may agree that it happens frequently enough:

          • AI generated articles
          • irrelevant SEO results
          • ads/sponsored results/commercial products or services
          • blog spam by people who speak out of ignorance
          • flame bait
          • deliberate disinformation
          • low-quality journalism
          • websites designed to exploit people/optimised for purposes other than to contribute to a healthy internet

          etc.

          • Arkive
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            6 months ago

            Thanks for your post. You’ve actually somewhat brought me around on AI search with your perplexity example. My previous AI search experiences have been general LLMs like ChatGPT (Opaque source data means I have to verify with traditional web search anyways) and Google’s new AI search feature (I’m uncomfortable with google discouraging traffic to the broader web). Since perplexity actually shows and links its sources, I’m going to give a try for the next few days alongside my usual DDG searches.

            I would be interested if you have an example of a search with mostly malicious results, since your stated experience seems disproportionate to my own. While I do concur that some results/websites are antagonistic towards my goal of useful information, I’m quite surprised to see someone say that they hate visiting websites in general. (Perhaps I’m missing hyperbole?)

            A bit of a digression, but it amused me to see you say you struggle to word your query for search engines, because I’ve typically had more problems wording my query for LLMs. I wonder if this is could be attributed to communication preferences, or just due to me having used search engines for almost 2 decades.