this post was submitted on 16 Jun 2024
293 points (98.7% liked)

Asklemmy

43942 readers
534 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] saltesc@lemmy.world 40 points 5 months ago (3 children)

I am so over hearing about AI. It's getting to the point that I can assume anyone dropping the term at work is an idiot that hasn't actually used or utilised it.

It's this LLM phase. It's super cool and a big jump in AI, but it's honestly not that good. It's a handy tool and one you need to heavily scrutinise beyond basic tasks. Businesses that jumped on it are now seeing the negative effects of thinking it was magic from the future that does everything. The truth is, it's stupid and people need to learn about it, understand it, and be trained in how to use it before it can be effective. It is a tool, not a solution—at least for now anyways.

[–] liam070@sopuli.xyz 24 points 5 months ago* (last edited 5 months ago) (2 children)

The truth is, it’s stupid and people need to learn about it, understand it, and be trained in how to use it before it can be effective.

So, like a hammer. A very expensive, environment-destroying hammer.

[–] saltesc@lemmy.world 5 points 5 months ago* (last edited 5 months ago)

That's actually a pretty good analogy.

I think more like discovering making fire or something. 90% of all the energy burnt is people worshipping it as it blazes away, never actually fulfilling any practical use except being marvelous to be around.

But once the forest is all chopped down, people are forced to understand fire and realise a couple small logs in a contained place was all they needed to have it be incredibly effective.

Oh, but that's too hard. It's magic right now. All hail the AI bonfire!

[–] notfromhere@lemmy.ml 2 points 5 months ago (1 children)

Genuinely curious, how does it destroy the environment?

[–] liam070@sopuli.xyz 11 points 5 months ago (3 children)

Massive energy consumption. Huge datacenters and not enough green energy. Now they want to build small nuclear plants. Without talking about the waste problem.

[–] bountygiver@lemmy.ml 2 points 5 months ago

Their waste is less destructive than coal plant though. Perhaps this could be a silver lining to finally get nuclear back in action and get closer to dropping coal once and for all.

[–] TheWeirdestCunt@lemm.ee 1 points 5 months ago

Tbf the energy issues are getting better, or at least there are some more efficient models being created. Back in April there was a version of Llama that only needed 8gb to almost match GPT4

[–] notfromhere@lemmy.ml 1 points 5 months ago (1 children)

So AI uses energy, and it’s how we are choosing to provide that energy is destructive to the environment? So AI isn’t itself destructive.

[–] oo1@lemmings.world 2 points 5 months ago (1 children)

Ah yeah, just choose a different energy souce. Simples.

Have you seen the growth in % of renewable (incl, nuc biofuel and waste) electricity generation over the past 30 years. (36% i in 1990 , dropped to about 33% in late 2000s up to 38% recently) this is global, IEA figures.

There have been two years since 1990 when renewable electricity output has grown faster than total electricity demand. 2008/9 recession and 2020 covid. The only way renewables will come close to meeting current electricity consumption is actually to start reducing those demands.

If we start transerffing gas( domestic heating), and petrol( low-capacity road transportation) onto the electricitry grid then the scale and speed of renewables needs to ramp up inconcievably quickly - it has grown fast over the past decade, but it hasn't been cheap nor has it been fast enough to keep up with current demands.

TBH I don't know where AI lines up next to EVs in scale of potential extra demand, probably lower but still an added demand (unless it can substitute for other stuff and improve efficiency somehow).

Electricity source is not really a choice, it is resource and tech constrained many sources are needed; the cheapest fuels will continue to be in the mix used so long as demand keeps increasing so fast.

Maybe, If you ran all AI in peooles houses in cold countries in winter, it'd substitute for heating - that'd be one way it could reduce its impact. Or maybe it can get its act together and spark widespread, frequent, deep, long lasting recessions in economic activity.

[–] notfromhere@lemmy.ml 0 points 5 months ago (1 children)

Maybe renewables is not the solution to our energy needs if it can’t scale up like we thought it could. Conservation of energy is not the answer. We as a society must find new, cleaner, sources of energy. Maybe AI can help us do it.

[–] oo1@lemmings.world 2 points 5 months ago

Perpetual motion machines are one of the mothers of all snake oil. Maybe AI can turn base metals into gold too. Do these AIs even really have a demonstrable understanding of thermodynamics yet? It needs to prove itself with a usable output for a clear observabe application on a small scale scale before anyone should start chucking vast amounts of energy at it in hope of what it can "maybe" do. I'd much rather chuck all that energy into trials of tokamaks or something like that.

[–] vrek@programming.dev 16 points 5 months ago

I equate an AI to an intern. It's useful for some stuff but if I'm going to attach my name to it I'm going to review it and probably change a lot about it.

[–] freebee@sh.itjust.works 2 points 5 months ago (1 children)

There's one good use case for me: produce a bigload of trialcontent in no time for load testing new stuff. "Make 2000 yada yada with column x and z ...". Keeps testing fun and varied while lots of testdata and that it's all nonsense doesn't matter.

I've found that testing code or formulas with LLM is a 50/50 now. Very often replying "use function blabla() and such snd so" very detailed instructions while this suggested function just doesn't exist at all in certain language asked for... it's still something I'ld try if I'm very stuck tho, never know.

[–] saltesc@lemmy.world 1 points 5 months ago

Very often replying "use function blabla() and such snd so" very detailed instructions while this suggested function just doesn't exist at all in certain language asked fo

I've noticed this a lot too—especially for M. But even though it makes up a function, it sometimes inspires a more optimised idea/method that can be more flexible for future datasets.

But most times it starts to massacre things and disregard prompted parameters or even producing an identical suggestion immediately after being told not to, why not to, and reconfirming original parameters of the query.

Some times punching in the same prompts five times for five iterations produces completely different results, but one may be on the right track and I can code the rest. It helps to set it's personality first, so it's sharing ideas it's seen out there, rather than trying to please.

At the least, it's a big time saver. Gone are the days where I get a few days spare to work on solving a complex problem through trial and discovery, so it's an excellent tool for reducing testing time and speeding up the route to an optimised method.