2
submitted 4 months ago by misk@sopuli.xyz to c/technology@lemmy.world
all 44 comments
sorted by: hot top controversial new old
[-] ShinkanTrain@lemmy.ml 2 points 4 months ago* (last edited 4 months ago)

Ok.

> uses search engine

> search engine gives generative AI answer

God dammit

> scroll down

> click search result

> AI Generated article

[-] RobotToaster@mander.xyz 2 points 4 months ago

> search engine gives generative AI answer

> It cites it source, so can't be that bad right?

> click link to source

> It's an AI generated article

Oh no.

[-] RisingSwell@lemmy.dbzer0.com 2 points 4 months ago

AI will give the correct, real source and then still make shit up. Bing linked to bulbapedia to tell me wailord was the heaviest Pokemon. Bulbapedia knows it isn't close, bingpt doesn't know shit.

[-] Matriks404@lemmy.world 0 points 4 months ago

It's funny because I've also used LLM for getting useful info about pokemon, and it didn't make any sense.

[-] prettybunnys@sh.itjust.works 2 points 4 months ago

It’s fantastic at templating

Just don’t trust what it provides the template

[-] FlashMobOfOne@lemmy.world 0 points 4 months ago
[-] Captainvaqina@sh.itjust.works 0 points 4 months ago
[-] FlashMobOfOne@lemmy.world 1 points 4 months ago

There's also udm14.com if you want to have cheeky fun with it.

[-] BananaTrifleViolin@lemmy.world 2 points 4 months ago

The Internet was a great resource for sharing and pooling human knowledge.

Now generative AI has come along to dilute knowledge in a great sea of excrement. Humans have to hunt through the shit to find knowledge.

[-] GaiusBaltar@lemmy.world 1 points 4 months ago

To be fair, humans were already diluting it in a great sea of excrement, the robots just came to take our job and do it even faster and better.

[-] Agent641@lemmy.world 1 points 4 months ago

When search engines stop being shit, I will.

[-] curiousaur@reddthat.com 0 points 4 months ago

Who else is going to aggregate those recipes for me without having to scroll past ads a personal blog bs?

[-] Nalivai@lemmy.world 0 points 4 months ago
[-] curiousaur@reddthat.com 0 points 4 months ago

Tell me you're not using them without telling me you're not using them.

[-] Nalivai@lemmy.world 0 points 4 months ago

Thd fuck do you mean without telling? I am very explicitly telling you that I don't use them, and I'm very openly telling you that you also shouldn't

[-] curiousaur@reddthat.com 0 points 4 months ago

I use them hundreds of times daily. I'm 3-5x more productive thanks to them. I'm incorporating them into the products I'm building to help make others who use the platform more productive.

Why the heck should I not use them? They are an excellent tool for so many tasks, and if you don't stay on top of their use, in many fields you will fall irrecoverably behind.

[-] rdsm@discuss.tchncs.de 0 points 4 months ago
[-] Oka@sopuli.xyz 0 points 4 months ago

I ask GPT for random junk all the time. If it's important, I'll double-check the results. I take any response with a grain of salt, though.

[-] zarkanian@sh.itjust.works 0 points 4 months ago

So, if it isn't important, you just want an answer, and you don't care whether it's correct or not?

[-] 0oWow@lemmy.world 0 points 4 months ago

The same can be said about the search results. For search results, you have to use your brain to determine what is correct and what is not. Now imagine for a moment if you were to use those same brain cells to determine if the AI needs a check.

AI is just another way to process the search results, that happens to give you the correct answer up front, most of the time. If you go blindly trust it, that's on you.

[-] zarkanian@sh.itjust.works -1 points 4 months ago

With the search results, you know what the sources are. With AI, you don't.

[-] Nalivai@lemmy.world 0 points 4 months ago

You are spending more time and effort doing that than you would googling old fashioned way. And if you don't check, you might as well throwing magic 8-ball, less damage to the environment, same accuracy

[-] bradd@lemmy.world 0 points 4 months ago

When it's important you can have an LLM query a search engine and read/summarize the top n results. It's actually pretty good, it'll give direct quotes, citations, etc.

[-] Nalivai@lemmy.world 0 points 4 months ago

And some of those citations and quotes will be completely false and randomly generated, but they will sound very believable, so you don't know truth from random fiction until you check every single one of them. At which point you should ask yourself why did you add unneccessary step of burning small portion of the rainforest to ask random word generator for stuff, when you could not do that and look for sources directly, saving that much time and energy

[-] PapstJL4U@lemmy.world 2 points 4 months ago

I, too, get the feeling, that the RoI is not there with LLM. Being able to include "site:" or "ext:" are more efficient.

I just made another test: Kaba, just googling kaba gets you a german wiki article, explaining it means KAkao + BAnana

chatgpt: It is the combination of the first syllables of KAkao and BEutel - Beutel is bag in german.

It just made up the important part. On top of chatgpt says Kaba is a famous product in many countries, I am sure it is not.

[-] bradd@lemmy.world 0 points 4 months ago

I guess it depends on your models and tool chain. I don't have this issue but I have seen it for sure, in the past with smaller models no tools and legal code.

[-] Nalivai@lemmy.world 1 points 4 months ago

You do have this issue, you can't not have this issue, your LLM, no matter how big the model is and how much tooling you use, does not have criteria for truth. The fact that you made this invisible for you is worse, so much worse.

[-] bradd@lemmy.world 1 points 4 months ago

If I put text into a box and out comes something useful I could give a shit less if it has a criteria for truth. LLM's are a tool, like a mannequin, you can put clothes on it without thinking it's a person, but you don't seem to understand that.

I work in IT, I can write a bash script to set up a server pivot to an LLM and ask for a dockerfile that does the same thing, and it gets me very close. Sure, I need to read over it and make changes but that's just how it works in the tech world. You take something that someone wrote and read over it and make changes to fit your use case, sometimes you find that real people make really stupid mistakes, sometimes college educated people write trash software, and that's a waste of time to look at and adapt... much like working with an LLM. No matter what you're doing, buddy, you still have to use your brian.

[-] bradd@lemmy.world 0 points 4 months ago

As a side note, I feel like this take is intellectually lazy. A knife cannot be used or handled like a spoon because it's not a spoon. That doesn't mean the knife is bad, in fact knives are very good, but they do require more attention and care. LLMs are great at cutting through noise to get you closer to what is contextually relevant, but it's not a search engine so, like with a knife, you have to be keenly aware of the sharp end when you use it.

[-] Nalivai@lemmy.world 1 points 4 months ago* (last edited 4 months ago)

LLMs are great at cutting through noise

Even that is not true. It doesn't have aforementioned criteria for truth, you can't make it have one.
LLMs are great at generating noise that humans have hard time distinguishing from a text. Nothing else. There are indeed applications for it, but due to human nature, people think that since the text looks like something coherent, information contained will also be reliable, which is very, very dangerous.

[-] bradd@lemmy.world 1 points 4 months ago

I understand your skepticism, but I think you're overstating the limitations of LLMs. While it's true that they can generate convincing-sounding text that may not always be accurate, this doesn't mean they're only good at producing noise. In fact, many studies have shown that LLMs can be highly effective at retrieving relevant information and generating text that is contextually relevant, even if not always 100% accurate.

The key point I was making earlier is that LLMs require a different set of skills and critical thinking to use effectively, just like a knife requires more care and attention than a spoon. This doesn't mean they're inherently 'dangerous' or only capable of producing noise. Rather, it means that users need to be aware of their strengths and limitations, and use them in conjunction with other tools and critical evaluation techniques to get the most out of them.

It's also worth noting that search engines are not immune to returning inaccurate or misleading information either. The difference is that we've learned to use search engines critically, evaluating sources and cross-checking information to verify accuracy. We need to develop similar critical thinking skills when using LLMs, rather than simply dismissing them as 'noise generators'.

See these:

[-] Oka@sopuli.xyz 0 points 4 months ago

The latest GPT does search the internet to generate a response, so it's currently a middleman to a search engine.

[-] Nalivai@lemmy.world 0 points 4 months ago

No it doesn't. It incorporates unknown number of words from the internet into a machine which only purpose is to sound like a human. It's an insanely complicated machine, but the truthfulness of the response not only never considered, but also is impossible to take as a deaired result.
And the fact that so many people aren't equipped to recognise it behind the way it talks could be buffling, but also very consistent with other choices humanity takes regularly.

[-] curiousaur@reddthat.com -1 points 4 months ago
[-] _cryptagion@lemmy.dbzer0.com -1 points 4 months ago
[-] leanleft@lemmy.ml 0 points 4 months ago

searX still uses the same search engines.

[-] _cryptagion@lemmy.dbzer0.com -1 points 4 months ago

Yes, however, using a public SearXNG instance makes your searches effectively private, since it’s the server doing them, not you. It also does not use generative AI to produce the results, and won’t until or unless the ability for normal searches is removed.

And at that point, you can just disable that engine for searching.

[-] leanleft@lemmy.ml 0 points 4 months ago

from a privacy perspective..
you might as well use a vpn or tor. same thing.

[-] _cryptagion@lemmy.dbzer0.com 0 points 4 months ago

Yes, but that’s not the only benefit to it. It’s a metasearch engine, meaning it searches all the individual sites you ask for, and combines the results into one page. This makes it more akin to DDG, but it doesn’t just use one search provider.

[-] leanleft@lemmy.ml 2 points 4 months ago

it's a fantastic metasearch engine. but also people frequently dont configure it to its max potential IMO . one common mishap is the frequent default setting of sending queries to google. 💩

[-] bradd@lemmy.world 0 points 4 months ago* (last edited 4 months ago)

I legiterally have an LLM use searxng for me.

[-] _cryptagion@lemmy.dbzer0.com -1 points 4 months ago

brother eww

this post was submitted on 05 Dec 2024
2 points (75.0% liked)

Technology

69351 readers
1784 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS