• Showroom7561@lemmy.ca
    link
    fedilink
    arrow-up
    12
    ·
    4 days ago

    AI LLMs have been pretty shit, but the advancement in voice, image generation, and video generation in the last two years has been unbelievable.

    We went from the infamous Will Smith eating spaghetti to videos that are convincing enough to fool most people… and it only took 2-3 years to get there.

    But LLMs will have a long way to go because of how they create content. It’s very easy to poison LLM datasets, and they get worse learning from other generated content.

    • MiyamotoKnows@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 day ago

      Poisoning LLM datasets is fun and easy! Especially when our online intellectual property is scraped (read: stolen) during training and no one is being accountable for it. Fight back! It’s as easy as typing false stuff at the end of your comments. As an 88 year old ex-pitcher for the Yankees who just set the new world record for catfish noodling you can take it from me!

  • ipkpjersi@lemmy.ml
    link
    fedilink
    arrow-up
    25
    arrow-down
    3
    ·
    4 days ago

    I’d argue it has. Things like ChatGPT shouldn’t be possible, maybe it’s unpopular to admit it but as someone who has been programming for over a decade, it’s amazing that LLMs and “AI” has come as far as it has over the past 5 years.

    That doesn’t mean we have AGI of course, and we may never have AGI, but it’s really impressive what has been done so far IMO.

    • jacksilver@lemmy.world
      link
      fedilink
      arrow-up
      9
      ·
      4 days ago

      If you’ve been paying attention to the field, you’d see it’s been a slow steady march. The technology that LLMs are based in were first published in 2016/2017, ChatGPT was the third iteration of the same base model.

      Thats not even accounting for all the work done with RNNs and LSTMs prior to that, and even more prior.

      Its definitely a major breakthrough, and very similar to what CNNs did for computer vision further back. But like computer vision, advancements have been made in other areas (like the generative space) and haven’t followed a linear path of progress.

    • Tedesche@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      4 days ago

      Agreed. I never thought it would happen in my lifetime, but it looks like we’re going to have Star Trek computers pretty soon.

  • blarghly@lemmy.world
    link
    fedilink
    English
    arrow-up
    37
    ·
    5 days ago

    When people talk about AI taking off exponentially, usually they are talking about the AI using its intelligence to make intelligence-enhancing modifications to itself. We are very much not there yet, and need human coaching most of the way.

    At the same time, no technology ever really follows a particular trend line. It advances in starts and stops with the ebbs and flows of interest, funding, novel ideas, and the discovered limits of nature. We can try to make projections - but these are very often very wrong, because the thing about the future is that it hasn’t happened yet.

    • haui@lemmy.giftedmc.com
      link
      fedilink
      arrow-up
      6
      arrow-down
      6
      ·
      5 days ago

      Although i agree with the general idea, AI (as in llms) is a pipe dream. Its a non product, another digital product that hypes investors up and produces “value” instead of value.

      • kescusay@lemmy.world
        link
        fedilink
        arrow-up
        10
        ·
        5 days ago

        Not true. Not entirely false, but not true.

        Large language models have their legitimate uses. I’m currently in the middle of a project I’m building with assistance from Copilot for VS Code, for example.

        The problem is that people think LLMs are actual AI. They’re not.

        My favorite example - and the reason I often cite for why companies that try to fire all their developers are run by idiots - is the capacity for joined up thinking.

        Consider these two facts:

        1. Humans are mammals.
        2. Humans build dams.

        Those two facts are unrelated except insofar as both involve humans, but if I were to say “Can you list all the dam-building mammals for me,” you would first think of beavers, then - given a moment’s thought - could accurately answer that humans do as well.

        Here’s how it goes with Gemini right now:

        Now Gemini clearly has the information that humans are mammals somewhere in its model. It also clearly has the information that humans build dams somewhere in its model. But it has no means of joining those two tidbits together.

        Some LLMs do better on this simple test of joined-up thinking, and worse on other similar tests. It’s kind of a crapshoot, and doesn’t instill confidence that LLMs are up for the task of complex thought.

        And of course, the information-scraping bots that feed LLMs like Gemini and ChatGPT will find conversations like this one, and update their models accordingly. In a few months, Gemini will probably include humans in its list. But that’s not a sign of being able to engage in novel joined-up thinking, it’s just an increase in the size and complexity of the dataset.

        • Zexks@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          4 days ago

          The biggest problem with LLMs as most currently use them is their inability to mull things over. To have multiple trains of thought and then try to intersect them and fork compilations of thought. When you ask a question it has exactly one chance to think up a response and no chance to review that thought or reconsider it. There are models that are allowed to do this but they’re generally behind pay walls because even the simplest of questions can lead into ridiculous tangents without proper guidelines on the prompt. Here the ‘Advanced resoning’ models response to the same question

          Mammals known to build dams

          #Mammal (scientific name)Dam-building habitKey reference1North American beaver (Castor canadensis)Constructs multi-year stick-and-mud dams on streams and ditches to flood an area deep enough for its lodge and food cache.2Eurasian beaver (Castor fiber)Same engineering instinct as its North-American cousin; creates extensive pond systems across Europe and parts of Asia.3Humans (Homo sapiens)From earthen farm ponds to megaprojects such as Hoover Dam, people build dams for water storage, flood control, power and more.

          Why the list is so short

          Beavers are unique. Despite a variety of lodge-building or burrowing rodents (muskrats, nutria, water voles, rakali, etc.), none of them actually dam a watercourse; they rely on natural water levels or on beaver-made ponds.

          No other living mammal species has been documented creating intentional water-blocking structures. (The extinct giant beaver Castoroides probably did not dam rivers, according to paleontological evidence. )

          So, when it comes to true dam-building in the mammal world, it’s essentially a two-species beaver monopoly—plus us.

          https://chatgpt.com/share/683caddc-5944-8009-8e4a-d03bef5933a4

          Also note that this response took a considerable amount more time than a standard response because it keeps reviewing it’s responses. But it’s worthwhile watching it’s thought process as it builds your answers.

        • haui@lemmy.giftedmc.com
          link
          fedilink
          arrow-up
          2
          arrow-down
          3
          ·
          5 days ago

          We’ll have to agree to disagree then. My hype argument perfectly matches your point of people wrongly perceiving llms as ai but my point goes further.

          AI is a search engine on steroids with all the drawbacks. It produces no more accurate results, has no more information, does not do anything else but take the research effort away which is proven to make people dumber. More importantly, llms gobble up energy like crazy and need rare ressources which are taken from exploited countries. In addition to that, they are a privacy nightmare and proven to systematically harm small creators due to breach of intellectual property, which is especially brutal for them.

          So no, there is no redeeming qualities in llms in their current form. They should be outlawed immediately and at most, locally used in specific cases.

    • Clinicallydepressedpoochie@lemmy.worldOP
      link
      fedilink
      arrow-up
      2
      arrow-down
      5
      ·
      edit-2
      5 days ago

      I do expect advancement to hit a period of exponential growth that quickly surpasses human intelligence. Given it adapts the drive to autonmously advance. Whether that is possible is yet to be seen and that’s kinda my point.

        • Zexks@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          edit-2
          4 days ago

          No “they” haven’t unless you can cite your source. Chatgpt was only released 2.5 years ago and even openai was saying 5-10 years with most outside watchers saying 10-15 with real nay sayers going out to 25 or more

        • Zexks@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          4 days ago

          Here are all 27 U.S. states whose names contain the letter “o”:

          Arizona

          California

          Colorado

          Connecticut

          Florida

          Georgia

          Idaho

          Illinois

          Iowa

          Louisiana

          Minnesota

          Missouri

          Montana

          New Mexico

          New York

          North Carolina

          North Dakota

          Ohio

          Oklahoma

          Oregon

          Rhode Island

          South Carolina

          South Dakota

          Vermont

          Washington

          Wisconsin

          Wyoming

          (That’s 27 states in total.)

          What’s missing?

  • mxeff@feddit.org
    link
    fedilink
    arrow-up
    53
    arrow-down
    5
    ·
    5 days ago

    This is precisely a property of exponential growth, that it can take (seemingly) very long until it starts exploding.

    • Clinicallydepressedpoochie@lemmy.worldOP
      link
      fedilink
      arrow-up
      6
      arrow-down
      23
      ·
      edit-2
      5 days ago

      What are you talking about it asymptoped at 5 units. It cant be described as exponential until it is exponential otherwise its better described as linear or polynomial if you must.

      • mxeff@feddit.org
        link
        fedilink
        arrow-up
        48
        ·
        edit-2
        5 days ago

        Exponential growth is always exponential, not just if it suddenly starts to drastically increase in the arbitrarily choosen view scale.

        A simple way, to check wether data is exponential, is to visualize it in loc-scale, and if it shows there a linear behavior, it has a exponential relation.

        Exponential growth means, that the values change by a constant ratio, contrary to linear growth where the data changes by a constant rate.

          • agamemnonymous@sh.itjust.works
            link
            fedilink
            arrow-up
            14
            ·
            5 days ago

            No, exponential functions are that way. A feature of exponential functions is that it increases very slowly until the slope hits 1. We’re still on the slow part, we didn’t really have any way of knowing exactly the extreme increase will be.

            • NaNin@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              3
              ·
              5 days ago

              Do you think that our current iteration of A.I. can have these kinds if gains? Like, what if the extreme increase happens beyond our lifetimes? or beyond the lifetime of our planet?

              • agamemnonymous@sh.itjust.works
                link
                fedilink
                arrow-up
                4
                ·
                5 days ago

                I think we can’t know, but LLMs definitely feel like a notable acceleration. Exponential functions are also, well, exponential. As X grows, X × X grows faster. The exponential part is gonna come from meta-models, coordinating multiple specialized models to complete complex tasks. Once we get a powerful meta-model, we’re off to the races. AI models developing AI models.

                It could take 50 years, it could take 5, it could happen this Wednesday. We won’t know which development is going to be the one to tip us over the edge until it happens, and even then only in retrospect. But it could very well be soon.

              • leftzero@lemmynsfw.com
                link
                fedilink
                arrow-up
                5
                arrow-down
                3
                ·
                5 days ago

                No, LLMs have always been an evident dead end when it comes to general AI.

                They’re hampering research in actual AI, and the fact that they’re being marketed as AI ensures that no one will invest in actual AI research in decades after the bubble bursts.

                We were on track for a technological singularity in our lifetimes, until those greedy bastards derailed us and murdered the future by poisoning the Internet with their slop for some short term profits.

                Now we’ll go extinct due to ignorance and global warming long before we have time to invent something smart enough to save us.

                But, hey, at least, for a little while, their line did go up, and that’s all that matters, it seems.

          • ArbitraryValue@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            10
            ·
            edit-2
            5 days ago

            An exponential function is a precise mathematical concept, like a circle or an even number. I’m not sure what you mean by “asymptote” here - an exponential function of the form y = k^x asymptotically approaches zero as x goes to negative infinity, but that doesn’t sound like what you’re referring to.

            People often have bad intuition about how exponential functions behave. They look like they grow slowly at first but that doesn’t mean that they’re not growing exponentially. Consider the story about the grains of rice on a chessboard.

                • Clinicallydepressedpoochie@lemmy.worldOP
                  link
                  fedilink
                  arrow-up
                  2
                  arrow-down
                  9
                  ·
                  edit-2
                  5 days ago

                  Tell me how im wrong. Or why did you even bother?

                  Or you can just admit you dont have any data to quantify your assertion that AI advancement is exponential growth. So youre just going off vibes.

                  Would you even admit that linear growth can grow faster than exponential growth?

                  Edit:

                  How about this, this is a real easy one.

                  What type of function is this:

              • calcopiritus@lemmy.world
                link
                fedilink
                arrow-up
                6
                ·
                5 days ago

                The exponential function has a single horizontal asymptote at y=0. Asymptotes at x=1 and x=-4 would be vertical. Exponential functions have no vertical asymptotes.

      • magic_lobster_party@fedia.io
        link
        fedilink
        arrow-up
        9
        arrow-down
        1
        ·
        5 days ago

        The derivative of an exponential is exponential. The relative difference between -1 and -2 is the same as 1 and 2.

        I’d say the development is exponential. Compare what we had 4 years ago, 2 years ago and now. 4 years ago it was inconceivable that an AI model could generate any convincing video at all. 2 years ago we laughed at Will Smith eating pasta. Today we have Veo 3 which generates videos with sound that are near indistinguishable from real life.

        It’s not going to be long until you regularly see AI generated videos without realizing it’s AI.

  • Pulptastic@midwest.social
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    4 days ago

    It has slowed exponentially because the models get exponentially more complicated the more you expect it to do.

    • linearchaos@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 days ago

      The exponential problem has always been there. We keep finding tricks and optimizations in hardware and software to get by it but they’re only occasional.

      The pruned models keep getting better so now You’re seeing them running on local hardware and cell phones and crap like that.

      I don’t think they’re out of tricks yet, but God knows when we’ll see the next advance. And I don’t think there’s anything that’ll take this current path into AGI I think that’s going to be something else.

  • chicken@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    26
    arrow-down
    1
    ·
    5 days ago

    A few years ago I remember people being amazed that prompts like “Markiplier drinking a glass of milk” could give them some blobs that looked vaguely like the thing asked for occasionally. Now there is near photorealistic video output. Same kind of deal with ability to write correct computer code and answer questions. Most of the concrete predictions/bets people made along the lines of “AI will never be able to do ______” have been lost.

    What reason is there to think it’s not taking off, aside from bias or dislike of what’s happening? There are still flaws and limitations for what it can do, but I feel like you have to have your head in the sand to not acknowledge the crazy level of progress.

    • kescusay@lemmy.world
      link
      fedilink
      arrow-up
      10
      ·
      5 days ago

      It’s absolutely taking off in some areas. But there’s also an unsustainable bubble because AI of the large language model variety is being hyped like crazy for absolutely everything when there are plenty of things it’s not only not ready for yet, but that it fundamentally cannot do.

      You don’t have to dig very deeply to find reports of companies that tried to replace significant chunks of their workforces with AI, only to find out middle managers giving ChatGPT vague commands weren’t capable of replicating the work of someone who actually knows what they’re doing.

      That’s been particularly common with technology companies that moved very quickly to replace developers, and then ended up hiring them back because developers can think about the entire project and how it fits together, while AI can’t - and never will as long as the AI everyone’s using is built around large language models.

      Inevitably, being able to work with and use AI is going to be a job requirement in a lot of industries going forward. Software development is already changing to include a lot of work with Copilot. But any actual developer knows that you don’t just deploy whatever Copilot comes up with, because - let’s be blunt - it’s going to be very bad code. It won’t be DRY, it will be bloated, it will implement things in nonsensical ways, it will hallucinate… You use it as a starting point, and then sculpt it into shape.

      It will make you faster, especially as you get good at the emerging software development technique of “programming” the AI assistant via carefully structured commands.

      And there’s no doubt that this speed will result in some permanent job losses eventually. But AI is still leagues away from being able to perform the joined-up thinking that allows actual human developers to come up with those structured commands in the first place, as a lot of companies that tried to do away with humans have discovered.

      Every few years, something comes along that non-developers declare will replace developers. AI is the closest yet, but until it can do joined-up thinking, it’s still just a pipe-dream for MBAs.

      • chicken@lemmy.dbzer0.com
        link
        fedilink
        arrow-up
        1
        ·
        4 days ago

        But any actual developer knows that you don’t just deploy whatever Copilot comes up with, because - let’s be blunt - it’s going to be very bad code. It won’t be DRY, it will be bloated, it will implement things in nonsensical ways, it will hallucinate… You use it as a starting point, and then sculpt it into shape.

        Yeah, but I don’t know where you’re getting the “never will” or “fundamentally cannot do” from. LLMs used to be only useful for coding if you ask for simple self-contained functions in the most popular languages, and now we’re here; most requests with small scope, I’m getting a result that is better written than I could have done myself by spending way more time, it makes way fewer mistakes than before and can often correct them. That’s with only using local models which became actually viable for me less than a year ago. So why won’t it keep going?

        From what I can tell there is not very much actually standing in the way of sensible holistic consideration of a larger problem or codebase here, just context size limits and being more likely to forget things in the context window the longer it is, which afaik are problems being actively worked on where there’s no reason they would be guaranteed to remain unsolved. This also seems to be what is holding back agentic AI from being actually useful. If that stuff gets cracked, I think it’s going to mean things will start changing even faster.

    • CallateCoyote@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      4 days ago

      Agreed. LLM Ai has gotten insanely good insanely fast, and an LLM of course isn’t going to magically turn into an AGI. That’s a whole different ball game.

    • squaresinger@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      4 days ago

      Yes, the goal posts keep moving, but they do so for a rather solid reason: We humans are famously bad at understanding intelligence and at understanding the differences between human and computer intelligence.

      100 years ago, doing complex calculations was seen as something very complex that only reasonably smart humans could do. Computers could easily outcompete humans, because calculations are inherently easy for computers while very difficult for humans.

      30 years ago we thought that high-level chess was something reserved only to the smartest of humans, and that it was a decent benchmark for intelligence. Turns out, playing chess is something that benefits greatly from large memory and fast computations, so again, it was easy for computers while really hard for humans.

      Nowadays AI can do a lot of things we thought would be really hard to do, but that computers can actually do. But there’s hardly any task performed by LLMs where it’s actually better than a moderately proficient human being. (Apart from tasks like “Do homework task X”, where again LLMs benefit from large memory since they can just regurgitate stuff from the training set.)

    • Clinicallydepressedpoochie@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      arrow-down
      2
      ·
      5 days ago

      Linear growth can be faster than exponential growth. Exponential implys tomorrow we will see it advance faster then it did the day before so every day we would see even crazier shit.

  • moseschrute@lemmy.world
    link
    fedilink
    arrow-up
    14
    arrow-down
    1
    ·
    4 days ago

    It has taken off exponentially. It’s exponentially annoying that’s it’s being added to literally everything

  • conditional_soup@lemm.ee
    link
    fedilink
    arrow-up
    12
    ·
    5 days ago

    Well, the thing is that we’re hitting diminishing returns with current approaches. There’s a growing suspicion that LLMs simply won’t be able to bring us to AGI, but that they could be a part of or stepping stone to it. The quality of the outputs are pretty good for AI, and sometimes even just pretty good without the qualifier, but the only reason it’s being used so aggressively right now is that it’s being subsidized with investor money in the hopes that it will be too heavily adopted and too hard to walk away from by the time it’s time to start charging full price. I’m not seeing that. I work in comp sci, I use AI coding assistants and so do my co-workers. The general consensus is that it’s good for boilerplate and tests, but even that needs to be double checked and the AI gets it wrong a decent enough amount. If it actually involves real reasoning to satisfy requirements, the AI’s going to shit its pants. If we were paying the real cost of these coding assistants, there is NO WAY leadership would agree to pay for those licenses.

    • thru_dangers_untold@lemmy.world
      link
      fedilink
      arrow-up
      5
      ·
      4 days ago

      Yeah, I don’t think AGI = an advanced LLM. But I think it’s very likely that a transformer style LLM will be part of some future AGI. Just like human brains have different regions that can do different tasks, an LLM is probably the language part of the “AGI brain”.

    • Korhaka@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      edit-2
      4 days ago

      What are the “real costs” though? It’s free to run a half decent LLM locally on a mid tier gaming PC.

      Perhaps a bigger problem for the big AI companies rather then the open source approach.

      • conditional_soup@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        4 days ago

        Sure, but ChatGPT costs MONEY. Money to run, and MONEY to train, and then they still have to make money back for their investors after everything’s said and done. More than likely, the final tally is going to look like whole cents per token once those investor subsidies run out, and a lot of businesses are going to be looking to hire humans back quick and in a hurry.

  • Etterra@discuss.online
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    4 days ago

    How do you know it hasn’t and us just laying low? I for one welcome our benevolent and merciful machine overlord.

  • Xaphanos@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    3
    ·
    5 days ago

    A major bottleneck is power capacity. Is is very difficult to find 50Mwatts+ (sometime hundreds) of capacity available at any site. It has to be built out. That involves a lot of red tape, government contracts, large transformers, contractors, etc. the current backlog on new transformers at that scale is years. Even Google and Microsoft can’t build, so they come to my company for infrastructure - as we already have 400MW in use and triple that already on contract. Further, Nvidia only makes so many chips a month. You can’t install them faster than they make them.

      • themurphy@lemmy.ml
        link
        fedilink
        arrow-up
        6
        arrow-down
        1
        ·
        5 days ago

        And it’s pretty great at it.

        AI’s greatest use case is not LLM and people treat it like that because it’s the only thing we can relate to.

        AI is so much better and many other tasks.

        • justOnePersistentKbinPlease@fedia.io
          link
          fedilink
          arrow-up
          3
          ·
          4 days ago

          Need it to not exponentially degrade when AI content is fed in.

          Need creativity to be more than random chance deviations from the statistically average result in a mostly stolen dataset taken from actual humans.

      • moonking@lemy.lol
        link
        fedilink
        arrow-up
        25
        arrow-down
        21
        ·
        5 days ago

        Humans don’t actually think either, we’re just electricity jumping to nearby neural connections that formed based on repeated association. Add to that there’s no free will, and you start to see how “think” is a immeasurable metric.

      • daniskarma@lemmy.dbzer0.com
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        4 days ago

        Maybe we are statistical engines too.

        When I heard people talk they are also repeating the most common sentences that they heard elsewhere anyway.

  • utopiah@lemmy.world
    link
    fedilink
    arrow-up
    5
    ·
    4 days ago

    LOL… you did make me chuckle.

    Aren’t we 18months until developers get replaced by AI… for like few years now?

    Of course “AI” even loosely defined progressed a lot and it is genuinely impressive (even though the actual use case for most hype, i.e. LLM and GenAI, is mostly lazier search, more efficient spam&scam personalized text or impersonation) but exponential is not sustainable. It’s a marketing term to keep on fueling the hype.

    That’s despite so much resources, namely R&D and data centers, being poured in… and yet there is not “GPT5” or anything that most people use on a daily basis for anything “productive” except unreliable summarization or STT (which both had plenty of tools for decades).

    So… yeah, it’s a slow take off, as expected. shrug

  • nucleative@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    3
    ·
    5 days ago

    What do you consider having “taken off”?

    It’s been integrated with just about everything or is in the works. A lot of people still don’t like it, but that’s not an unusual phase of tech adoption.

    From where I sit I’m seeing it everywhere I look compared to last year or the year before where pretty much only the early adopters were actually using it.

    • capybara@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      3
      ·
      5 days ago

      What do you mean when you say AI has been integrated with everything? Very broad statement that’s obviously not literally true.

      • nucleative@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        5 days ago

        True, I tried to qualify it with just about or on the way.

        From the perspective of my desk, my core business apps have AI auto suggest in key fields (software IDEs, ad buying tools, marketing content preparation such as Canva). My Whatsapp and Facebook messenger apps now have an “Ask meta AI” feature front and center. Making a post on Instagram, it asks if I want AI assistance to write the caption.

        I use an app to track my sleeping rhythm and it has an AI sleep analysis feature built in. The photo gallery on my phone includes AI photo editing like background removal, editing things out (or in).

        That’s what I mean when I say it’s in just about everything, at least relative to where we were just a short bit of time ago.

        You’re definitely right that it’s not literally in everything.

        • kadup@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          4 days ago

          To be fair, smart background removal was a feature from Picasa over a decade ago. We just didn’t call everything “AI” to make shareholders happy.

  • neon_nova@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    2
    ·
    4 days ago

    I think we might not be seeing all the advancements as they are made.

    Google just showed off AI video with sound. You can use it if you subscribe to thier $250/month plan. That is quite expensive.

    But if you have strong enough hardware, you can generate your own without sound.

    I think that is a pretty huge advancement in the past year or so.

    I think that focus is being put on optimizing these current things and making small improvements to quality.

    Just give it a few years and you will not even need your webcam to be on. You could just use an AI avatar that look and sounds just like you running locally on your own computer. You could just type what you want to say or pass through audio. I think the tech to do this kind of stuff is basically there, it just needs to be refined and optimized. Computers in the coming years will offer more and more power to let you run this stuff.

  • Kyrgizion@lemmy.world
    link
    fedilink
    arrow-up
    18
    arrow-down
    3
    ·
    5 days ago

    It’s not anytime soon. It can get like 90% of the way there but those final 10% are the real bitch.

    • WhatAmLemmy@lemmy.world
      link
      fedilink
      English
      arrow-up
      45
      arrow-down
      1
      ·
      edit-2
      5 days ago

      The AI we know is missing the I. It does not understand anything. All it does is find patterns in 1’s and 0’s. It has no concept of anything but the 1’s and 0’s in its input data. It has no concept of correlation vs causation, that’s why it just hallucinates (presents erroneously illogical patterns) constantly.

      Turns out finding patterns in 1’s and 0’s can do some really cool shit, but it’s not intelligence.

      • Monstrosity@lemm.ee
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        4
        ·
        5 days ago

        This is not necessarily true. While it’s using pattern recognition on a surface level, we’re not entirely sure how AI comes up with it’s output.

        But beyond that, a lot of talk has been centered around a threshold when AI begins training other AI & can improve through iterations. Once that happens, people believe AI will not only improve extremely rapidly, but we will understand even less of what is happening when an AI black boxes train other AI black boxes.

        • Coldcell@sh.itjust.works
          link
          fedilink
          arrow-up
          2
          arrow-down
          1
          ·
          5 days ago

          I can’t quite wrap my head around this, these systems were coded, written by humans to call functions, assign weights, parse data. How do we not know what it’s doing?

          • MangoCats@feddit.it
            link
            fedilink
            English
            arrow-up
            3
            ·
            5 days ago

            It’s a bit of “emergent properties” - so many things are happening under the hood they don’t understand exactly how it’s doing what it’s doing, why one type of mesh performs better on a particular class of problems than another.

            The equations of the Lorenz attractor are simple, well studied, but it’s output is less than predictable and even those who study it are at a loss to explain “where it’s going to go next” with any precision.

          • Kyrgizion@lemmy.world
            link
            fedilink
            arrow-up
            4
            arrow-down
            1
            ·
            5 days ago

            Same way anesthesiology works. We don’t know. We know how to sedate people but we have no idea why it works. AI is much the same. That doesn’t mean it’s sentient yet but to call it merely a text predictor is also selling it short. It’s a black box under the hood.

            • Coldcell@sh.itjust.works
              link
              fedilink
              arrow-up
              3
              arrow-down
              3
              ·
              5 days ago

              Writing code to process data is absolutely not the same way anesthesiology works 😂 Comparing state specific logic bound systems to the messy biological processes of a nervous system is what gets this misattribution of ‘AI’ in the first place. Currently it is just glorified auto-correct working off statistical data about human language, I’m still not sure how a written program can have a voodoo spooky black box that does things we don’t understand as a core part of it.

              • irmoz@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                2
                ·
                edit-2
                5 days ago

                The uncertainty comes from reverse-engineering how a specific output relates to the prompt input. It uses extremely fuzzy logic to compute the answer to “What is the closest planet to the Sun?” We can’t know which nodes in the neural network were triggered or in what order, so we can’t precisely say how the answer was computed.

          • The_Decryptor@aussie.zone
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            2
            ·
            5 days ago

            Yeah, there’s a mysticism that’s sprung up around LLMs as if they’re some magic blackbox, rather than a well understood construct to the point where you can buy books from Amazon on how to write one from scratch.

            It’s not like ChatGPT or Claude appeared from nowhere, the people who built them do talks about them all the time.

            • Monstrosity@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              edit-2
              5 days ago

              What a load of horseshit lol

              EDIT: Sorry, I’ll expand. When AI researchers give talks about how AI works, they say things like, “on a fundamental level, we don’t actually know what’s going on.”

              Also, even if there are books available about how to write an AI from scratch(?) somehow, the basic understanding of what happens deep within the neural networks is still a “magic black box”. They’ll crack it open eventually, but not yet.

              The ideas that people have that AI is simple and stupid & a passing fad are naive.

              • The_Decryptor@aussie.zone
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 days ago

                If these AI researchers really have no idea how these things work, then how can they possibly improve the models or techniques?

                Like how they now claim all that after upgrades that now these LLMs can “reason” about problems, how did they actually go and add that if it’s a black box?

        • MangoCats@feddit.it
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          5 days ago

          Steam locomotive operators would notice some behaviors of their machines that they couldn’t entirely explain. They were out there, shoveling the coal, filling the boilers, and turning the valves but some aspects of how the engines performed - why they would run stronger in some circumstances than others - were a mystery to the men on the front lines. Decades later, intense theoretical study could explain most of the observed phenomena by things like local boiling inside the boiler insulating the surface against heat transfer from the firebox, etc. but at the time when the tech was new: it was just a mystery.

          Most of the “mysteries” of AI are similarly due to the fact that the operators are “vibe coding” - they go through the motions and they see what comes out. They’re focused on their objectives, the input-output transform, and most of them aren’t too caught up in the how and why of what it is doing.

          People will study the how and why, but like any new tech, their understanding is going to lag behind the actions of the doers who are out there breaking new ground.

      • MangoCats@feddit.it
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        2
        ·
        5 days ago

        Distill intelligence - what is it, really? Predicting what comes next based on… patterns. Patterns you learn in life, from experience, from books, from genetic memories, but that’s all your intelligence is too: pattern recognition / prediction.

        As massive as current AI systems are, consider that you have ~86 Billion neurons in your head, devices that evolved over the span of billions of years ultimately enabling you to survive in a competitive world with trillions of other living creatures, eating without being eaten at least long enough to reproduce, back and back and back for millions of generations.

        Current AI is a bunch of highly simplified computers with up to hundreds of thousands of cores. Like planes fly faster than birds, AI can do some tricks better than human brains, but mostly: not.

      • Pornacount128@lemmynsfw.com
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        5 days ago

        Humans are just nurons, we don’t “understand either” until so many stack on top of each other than we have a sort of consciousness. The it seems like we CAN understand but do we? Or are we just a bunch of meat computers? Also, llms handle language or correlations of words, don’t humans just do that (with maybe body language too) but we’re all just communicating. If llms can communicate isn’t that enough conceptually to do anything? If llms can program and talk to other llms what can’t they do?