• 0 Posts
  • 30 Comments
Joined 1 month ago
cake
Cake day: July 8th, 2025

help-circle









  • it seems dubious to make the claim that large language models write more efficient code. the popularity of node.js alone makes me doubt there’s all that much efficient code out there for it to train on, at least percentage-wise. i mean, the most popular app for hardcore gamers to run in the background packages and runs on its own copy of google chrome. add to that hallucinations and code quality and whatever else and i doubt its code is achieving the efficiency of a high school coding class, at least in the general case


  • Nobody is advocating for the service model companies like openai use here. I think this tech should be done using open source models that can be run locally.

    this is definitely fair. i think my big issue with it is the inordinate amount of capital (land, carbon emissions, water) that go into it. maybe i’ve unfairly associated all ai with openai and gemini and meta.

    Now compare that with DeepSeek.

    my understanding of deepseek is that most of their models are trained by engaging in dialogue with existing models. the cost of training and running those models should be taken into account in that case. if it is from scratch that might change things, if the carbon and water numbers are good.

    In the future, he said, 98% of AI applications in the market will serve industrial and agricultural needs while only 2% will serve consumers directly.

    i think that’s a problem with the definition of ai. it’s not clear to me what tim huawei defines ai as. i’m not arguing against the concept of machine learning, to be clear. i thought we were talking specifically about language models and photo and video generation and whatnot

    It’s easy to dismiss this stuff when you already have a bias against it and don’t want it to work, but the reality is that it’s already a useful tool once you learn where and when to use it.

    yeah that’s fair enough. i didn’t mean to get into a huge discussion over llms because there’s definitely an element of that in my head. idk, i guess my point in saying that was that you can shit out a more-or-less working piece of code in any language pretty quickly, if you don’t need it to be idiomatic or maintainable. my understanding was ai was kind of the same in that regard.

    i guess if training large language models can be done with negligible emissions and cooled with gray or black water, i can’t be against it. programming is definitely the main field where there’s no arguing that llms aren’t useful at all. i’m still unconvinced that’s what’s happening, even with deepseek, but if they’re putting their datacenters on 3-mile island and using sewage to cool their processors, i guess that would assuage my concerns.


  • LLMs get the brunt of it because alfalfa has more uses than chatgpt. maybe it’s the result of my own bias but i would consider golf courses more useful than chatgpt. LLMs aren’t even close to the sole contributor to climate change, but they are emblematic of venture capitalists more than i think anything else. but it’s hard for me to justify the creation and use of these things when they have very narrow use cases, often create as much work as they save, and suck down clean drinking water like i suck down whiskey sours


  • the IEA report was made in mid-2023, and i would imagine ai electricity usage has skyrocketed since then. as mentioned in the mit source, dating to may 2025, electricity usage by ai is 48% dirtier than the us average. my problem with ai isn’t that it violates intellectual property rights, it’s that llms are a net-negative to society because of their climate effects. if ai datacenters were built using clean energy and cooled using dirty water, it would likely be little more than a mild annoyance for me. as it stands, we are putting the global south underwater so that people who are surrounded by yes-men can have yes-robots too.


  • The efficiency has already improved dramatically

    the mit article was written this may, and as it notes, ai datacenters still use much more electricity than other datacenters, and that electricity is generated through less environmentally-friendly methods. openai, if it is solvent long enough to count, will

    build as many as 10 data centers (each of which could require five gigawatts, more than the total power demand from the state of New Hampshire)

    even the most efficient models take several orders of magnitude more energy to create than to use:

    it’s estimated that training OpenAI’s GPT-4 took over $100 million and consumed 50 gigawatt-hours of energy

    and overall, ai datacenters use

    millions of gallons of water (often fresh, potable water) per day

    i’m doubtful that the uses of llms justify the energy cost for training, especially when you consider that the speed at which they are attempting to create these “tools” requires that they use fossil fuels to do it. i’m not gonna make the argument that aggregate demand is growing, because i believe that the uses of llms are rather narrow, and if ai is being used more, it’s because it is being forced on the consumer in order for tech companies to post the growth numbers necessary to keep the line growing up. i know that i don’t want gemini giving me some inane answer every time i google something. maybe you do.

    if you use a pretrained model running locally, you know the energy costs of your queries better than me. if you use an online model running in a large datacenter, i’m sorry but doubting the environmental costs of making queries seems to be treatler cope more than anything else. even if you do use a pretrained model, the cost of creation likely eclipses the benefit to society of its existence.

    EDIT: to your first point, it takes a bit to learn how to write idiomatic code in a new paradigm. but if you’re super concerned about code quality you’re not using an llm anyway. at least unless they’ve made large strides since i last used one.


  • into_highest_invite@lemmygrad.mltoGenZedong@lemmygrad.mlMy thoughts on AI
    link
    fedilink
    arrow-up
    16
    arrow-down
    1
    ·
    edit-2
    7 days ago

    I would have to learn entire coding languages to do it myself, which takes years. AI can do it in 30 minutes and better than I could in years

    this is an overstatement. once you learn the basics of one programming language (which does not take a full year), you can apply the knowledge to other programming languages, many of which are almost identical to one another.

    There is of course the environmental cost. To that I want to say that everything has an environmental cost. I don’t necessarily deny AI is a water-hog, just that the way we go about it in capitalism, everything is contributing to climate change and droughts. Moreover to be honest I’ve never seen actual numbers and studies, everyone just says “generating this image emptied a whole bottle of water”. It’s just things people repeat idly like so many other things; and without facts, we cannot find truth.

    according to a commonly-cited 2023 study:

    training the GPT-3 language model in Microsoft’s state-of-the-art U.S. data centers can directly evaporate 700,000 liters of clean freshwater, but such information has been kept a secret

    the global AI demand is projected to account for 4.2 – 6.6 billion cubic meters of water withdrawal in 2027, which is more than the total annual water withdrawal of 4 – 6 Denmark or half of the United Kingdom.

    GPT-3 needs to “drink” (i.e., consume) a 500ml bottle of water for roughly 10 – 50 medium-length responses, depending on when and where it is deployed.

    there’s also the energy costs:

    according to google’s 2024 environmental report:

    In 2023, our total GHG emissions were 14.3 million tCO2e, representing a 13% year-over-year increase and a 48% increase compared to our 2019 target base year. This result was primarily due to increases in data center energy consumption and supply chain emissions. As we further integrate AI into our products, reducing emissions may be challenging due to increasing energy demands from the greater intensity of AI compute, and the emissions associated with the expected increases in our technical infrastructure investment.

    according to the mit technology review:

    The carbon intensity of electricity used by data centers was 48% higher than the US average.

    and

    [by 2028] AI alone could consume as much electricity annually as 22% of all US households.

    there’s also this article by the UN, but this comment is getting kinda long and the whole thing is relevant imo so it is left as an exercise to the reader

    i have my own biases against ai, so i’m not gonna try to write a full response, but this is what stood out to me






  • i haven’t gotten much sleep and it’s really early so apologies if i misunderstood the question. very broadly, there’s two kinds of open source licenses: copyleft and permissive. generally, permissive licenses like MIT allow any usage of the code, including by copyrighting your own contributions or including it in copyrighted works. copyleft licenses require additions to the code to be open sourced too. this was a problem for apple when GNU code updated from GPLv2 to v3, which iirc added the restriction that any package that included licensed programs also had to be copyleft. this was a problem because apple had packaged a lot of GPL programs with macOS, so now they haven’t been updated since 2007