361
submitted 3 months ago by Ilandar@aussie.zone to c/technology@lemmy.world
top 50 comments
sorted by: hot top controversial new old
[-] JiveTurkey@lemmy.world 84 points 3 months ago* (last edited 3 months ago)

Useful in the way that it increases emissions and hopefully leads to our demise because that's what we deserve for this stupid technology.

[-] Schal330@lemmy.world 64 points 3 months ago

Surely this is better than the crypto/NFT tech fad. At least there is some output from the generative AI that could be beneficial to the whole of humankind rather than lining a few people's pockets?

[-] JiveTurkey@lemmy.world 36 points 3 months ago

Unfortunately crypto is still somehow a thing. There is a couple year old bitcoin mining facility in my small town that brags about consuming 400MW of power to operate and they are solely owned by a Chinese company.

[-] EldritchFeminity@lemmy.blahaj.zone 7 points 3 months ago

I recently noticed a number of bitcoin ATMs that have cropped up where I live - mostly at gas stations and the like. I am a little concerned by it.

load more comments (9 replies)
load more comments (7 replies)
load more comments (9 replies)
[-] souperk@reddthat.com 5 points 3 months ago

While the consumption for AI train can be large, there are arguments to be made for its net effect in the long run.

The article's last section gives a few examples that are interesting to me from an environmental perspective. Using smaller problem-specific models can have a large effect in reducing AI emissions, since their relation to model size is not linear. AI assistance can indeed increase worker productivity, which does not necessarily decrease emissions but we have to keep in mind that our bodies are pretty inefficient meat bags. Last but not least, AI literacy can lead to better legislation and regulation.

[-] JiveTurkey@lemmy.world 13 points 3 months ago

The argument that our bodies are inefficient meat bags doesn't make sense. AI isn't replacing the inefficient meat bag unless I'm unaware of an AI killing people off and so far I've yet to see AI make any meaningful dent in overall emissions or research. A chatgpt query can use 10x more power than a regular Google search and there is no chance the result is 10x more useful. AI feels more like it's adding to the enshittification of the internet and because of its energy use the enshittification of our planet. IMO if these companies can't afford to build renewables to support their use then they can fuck off.

load more comments (1 replies)
[-] Wirlocke@lemmy.blahaj.zone 4 points 3 months ago* (last edited 3 months ago)

Theoretically we could slow down training and coast on fine-tuning existing models. Once the AI's trained they don't take that much energy to run.

Everyone was racing towards "bigger is better" because it worked up to GPT4, but word on the street is that raw training is giving diminishing returns so the massive spending on compute is just a waste now.

load more comments (2 replies)
[-] skittle07crusher@sh.itjust.works 26 points 3 months ago

Is it me or is there something very facile and dull about Gartner charts? Thinking especially about the “””magic””” quadrants one (wow, you ranked competitors in some area along TWO axes!), but even this chart feels like such a mundane observation that it seems like frankly undeserved advertising for Gartner, again, given how little it actually says.

[-] CosmoNova@lemmy.world 19 points 3 months ago

And it isn't even true in many cases. For example the internet with the dotcom bubble. It actually became much bigger and important than anyone anticipated in the 90s.

[-] Plopp@lemmy.world 14 points 3 months ago

The graph for VR would also be quite interesting, given how many hype cycles it has had over the decades.

load more comments (1 replies)
[-] Mr_Dr_Oink@lemmy.world 4 points 3 months ago

The trough of disillusionment sounds like my former depression. The slope of enlightenment sounds like a really frun water slide.

load more comments (8 replies)
[-] smackjack@lemmy.world 30 points 3 months ago

LLMs need to get better at saying "I don't know." I would rather an LLM admit that it doesn't know the answer instead of making up a bunch of bullshit and trying to convince me that it knows what it's talking about.

[-] Floey@lemm.ee 15 points 3 months ago

LLMs don't "know" anything. The true things they say are just as much bullshit as the falsehoods.

[-] EnderMB@lemmy.world 10 points 3 months ago

I work on LLM's for a big tech company. The misinformation on Lemmy is at best slightly disingenuous, and at worst people parroting falsehoods without knowing the facts. For that reason, take everything (even what I say) with a huge pinch of salt.

LLM's do NOT just parrot back falsehoods, otherwise the "best" model would just be the "best" data in the best fit. The best way to think about a LLM is as a huge conductor of data AND guiding expert services. The content is derived from trained data, but it will also hit hundreds of different services to get context, find real-time info, disambiguate, etc. A huge part of LLM work is getting your models to basically say "this feels right, but I need to find out more to be correct".

With that said, I think you're 100% right. Sadly, and I think I can speak for many companies here, knowing that you're right is hard to get right, and LLM's are probably right a lot in instances where the confidence in an answer is low. I would rather a LLM say "I can't verify this, but here is my best guess" or "here's a possible answer, let me go away and check".

[-] 31337@sh.itjust.works 5 points 3 months ago

I thought the tuning procedures, such as RLHF, kind of messes up the probabilities, so you can't really tell how confident the model is in the output (and I'm not sure how accurate these probabilities were in the first place)?

Also, it seems, at a certain point, the more context the models are given, the less accurate the output. A few times, I asked ChatGPT something, and it used its browsing functionality to look it up, and it was still wrong even though the sources were correct. But, when I disabled "browsing" so it would just use its internal model, it was correct.

It doesn't seem there are too many expert services tied to ChatGPT (I'm just using this as an example, because that's the one I use). There's obviously some kind of guardrail system for "safety," there's a search/browsing system (it shows you when it uses this), and there's a python interpreter. Of course, OpenAI is now very closed, so they may be hiding that it's using expert services (beyond the "experts" in the MOE model their speculated to be using).

load more comments (1 replies)
load more comments (1 replies)
[-] JDPoZ@lemmy.world 9 points 3 months ago* (last edited 3 months ago)

I hate to break this to everyone who thinks that “AI” (LLM) is some sort of actual approximation of intelligence, but in reality, it’s just a fucking fancy ass parrot.

Our current “AI” doesn’t understand anything or have context, it’s just really good at guessing how to say what we want it to say… essentially in the same way that a parrot says “Polly wanna cracker.”

A parrot “talking to you” doesn’t know that Polly refers to itself or that a cracker is a specific type of food you are describing to it. If you were to ask it, “which hand was holding the cracker…?” it wouldn’t be able to answer the question… because it doesn’t fucking know what a hand is… or even the concept of playing a game or what a “question” even is.

It just knows that it makes it mouth, go “blah blah blah” in a very specific way, a human is more likely to give it a tasty treat… so it mushes its mouth parts around until its squawk becomes a sound that will elicit such a reward from the human in front of it… which is similar to how LLM “training models” work.

Oversimplification, but that’s basically it… a trillion-dollar power-grid-straining parrot.

And just like a parrot - the concept of “I don’t know” isn’t a thing it comprehends… because it’s a dumb fucking parrot.

The only thing the tech is good at… is mimicking.

It can “trace the lines” of any existing artist in history, and even blend their works, which is indeed how artists learn initially… but an LLM has nothing that can “inspire” it to create the art… because it’s just tracing the lines like a child would their favorite comic book character. That’s not art. It’s mimicry.

It can be used to transform your own voice to make you sound like most celebrities almost perfectly… it can make the mouth noises, but has no idea what it’s actually saying… like the parrot.

You get it?

load more comments (1 replies)
load more comments (4 replies)
[-] IndiBrony@lemmy.world 24 points 3 months ago* (last edited 3 months ago)

To be fair, it is useful in some regards.

I'm not a huge fan of Amazon, but last time I had an issue with a parcel it was sorted out insanely fast by the AI assistant on the website.

Within literally 2 minutes I'd had a refund confirmed. No waiting for people to eventually pick up the phone after 40 minutes. No misunderstanding or annoying questions. The moment I pressed send on my message it instantly started formulating a reply.

The truncated version went:

"Hey I meant to get [x] delivery, but it hasn't arrived. Can I get a refund?"

"Sure, your money will go back into [y] account in a few days. If the parcel turns up in the meantime, you can send it back by dropping it off at [z]"

Done. Absolutely painless.

[-] jlh@lemmy.jlh.name 64 points 3 months ago

How is a chatbot here better, faster, or more accurate than just a "return this" button on a web page? Chat bots like that take 10x the programming effort and actively make the user experience worse.

[-] thetreesaysbark@sh.itjust.works 13 points 3 months ago

Presumably there could be nuance to the situation that the chat bot is able to convey?

[-] jlh@lemmy.jlh.name 8 points 3 months ago

But that nuance is probably limited to a paragraph or two of text. There's nothing the chatbot knows about the returns process at a specific company that isn't contained in that paragraph. The question is just whether that paragraph is shown directly to the user, or if it's filtered through an LLM first. The only thing I can think of is that chatbot might be able to rephrase things for confused users and help stop users from ignoring the instructions and going straight to human support.

load more comments (2 replies)
[-] Grandwolf319@sh.itjust.works 4 points 3 months ago

And it could hallucinate, so you would need to add further validation after the fact

[-] SkyNTP@lemmy.ml 29 points 3 months ago

That has nothing to do with AI and is strictly a return policy matter. You can get a return in less than 2 minutes by speaking to a human at Home Depot.

Businesses choose to either prioritize customer experience, or not.

load more comments (3 replies)
[-] KeefChief13@lemmy.world 5 points 3 months ago

I like using it to assist me when I am coding.

load more comments (5 replies)
load more comments (4 replies)
[-] AlexWIWA@lemmy.ml 19 points 3 months ago

Useful for scammers and spam

[-] Mwa@thelemmy.club 9 points 3 months ago

hot take: chatbots are actually kinda useful for problem solving but its not the best at it

load more comments (1 replies)
[-] yesman@lemmy.world 7 points 3 months ago

We should be using AI to pump the web with nonsense content that later AI will be trained on as an act of sabotage. I understand this is happening organically; that's great and will make it impossible to just filter out AI content and still get the amount of data they need.

[-] CosmoNova@lemmy.world 16 points 3 months ago

That sounds like dumping trash in the oceans so ships can't get through the trash islands easily anymore and become unable to transport more trashy goods. Kinda missing the forest for the trees here.

load more comments (1 replies)
[-] Toribor@corndog.social 7 points 3 months ago

My shitposting will make AI dumber all on its own; feedback loop not required.

[-] TheLadyAugust@lemmy.world 4 points 3 months ago

Alternatively, and possibly almost as useful, companies will end up training their AI to detect AI content so that they don't train on AI content. Which would in turn would give everyone a tool to filter out AI content. Personally, I really like the apps that poison images when they're uploaded to the internet.

[-] Zangoose@lemmy.world 4 points 3 months ago

Bold of you to assume companies will release their AI detection tools

load more comments (1 replies)
[-] Grandwolf319@sh.itjust.works 6 points 3 months ago

So if I were to get this straight, the entire logic is that due to big hype, it fits the pattern or other techs becoming useful… that’s sooo not a guarantee, so many big hype stuff have died.

load more comments
view more: next ›
this post was submitted on 20 Aug 2024
361 points (91.7% liked)

Technology

59583 readers
2575 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS