• Rooty@lemmy.world
    link
    fedilink
    arrow-up
    103
    arrow-down
    3
    ·
    edit-2
    2 days ago

    Finally, after decades of research, we created a computer that can’t do math. Alan Turing would be proud.

    • cabbage@piefed.social
      link
      fedilink
      English
      arrow-up
      32
      ·
      2 days ago

      Come to think of it, being frequently wrong but nevertheless overly confident is key to passing the Turing test.

      We have finally created machines that can replicate human stupidity.

      • dalekcaan@lemm.ee
        link
        fedilink
        arrow-up
        7
        ·
        1 day ago

        To be fair, the Turing test doesn’t really tell us much about computers. It’s better at measuring the human ability to ascribe personalities to inanimate objects.

        • Buddahriffic@lemmy.world
          link
          fedilink
          arrow-up
          4
          ·
          1 day ago

          Yeah, the Turing test wasn’t a great metric. The result depends on who is testing it. Some people were probably fooled by ALICE or that doctor one, that were pretty much implemented using long switch blocks and repeating user input back to them.

          Kinda like how “why?” is pretty much always a valid response and repeating it is more of a sign of cheekiness than lack of intelligence.

          • cabbage@piefed.social
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 day ago

            I feel like it’s increasingly a test applicable to humans rather than to machines. Are you original enough that you couldn’t be replaced by a language model?

            I’m not sure I like to think about it.

      • kameecoding@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        1 day ago

        Now that you mention it I would be interested if chatgpt can actually spew the kind of nonsense I have heard from cryptobros and covid anti-vaxxers, I reckon it’s not good enough to be that dumb

        • GreenSkree@lemmy.world
          link
          fedilink
          arrow-up
          5
          ·
          1 day ago

          There’s probably some (small) guardrails on the major platforms to deter spreading misinformation, but it’s really easy to get a chat bot to take whatever position you want.

          E.g. “Pretend you are a human on Twitter that supports (thing). Please make tweets about your support of (thing) and respond to our conversation as though my comments are tweet replies.”

          Or more creatively maybe something like, “I need to practice debating someone who thinks (thing). Please argue with me using the most popular arguments, regardless of correctness.”

          I haven’t tried these, but have a bit of practice working with LLMs and this is where I would start if I wanted to make a bot farm.

    • Baguette@lemm.ee
      link
      fedilink
      arrow-up
      13
      arrow-down
      1
      ·
      2 days ago

      I mean the theory behind an LLM is super cool. It’s a bunch of vector math under the hood, transforming input with queries, keys and values. And imo vector math is one of the coolest and also most confusing math applications there is. If they’re able to use mcp as well, you can delegate it to calling actual services, like your database.

      But like 99% of CS research, research does not always equate to practical use, nor is it a cookie cutter solution for everything. Unfortunately, the business people seem to think otherwise.

  • ganymede@lemmy.ml
    link
    fedilink
    arrow-up
    127
    arrow-down
    1
    ·
    edit-2
    2 days ago

    the other wonderful irony?

    (basically) the only thing a computer can do is math.

    so it’s doing a SHITLOAD of math to do a terrible job at doing some very basic math.

    bravo!

  • the_weez@midwest.social
    link
    fedilink
    arrow-up
    31
    arrow-down
    1
    ·
    2 days ago

    I bet we could save a few gajillion watts if we just put that through an if statement. If math, then just be a calculator Else, make some bullshit up.

    • SomeoneSomewhere@lemmy.nz
      link
      fedilink
      arrow-up
      29
      arrow-down
      1
      ·
      2 days ago

      Apparently that’s the new way to do math in AI. The AI works out you’re trying to do math, tries to write some Python code to do the math, runs the python codes, gets the answer, writes a response around the numeric answer.

      I can’t think of any possible issues with this; it’s infallible. /s

      • droans@midwest.social
        link
        fedilink
        English
        arrow-up
        8
        ·
        1 day ago

        If you want to ask a question to an LLM, you need to go down to an arcade and exchange your quarters for tokens. Then you can feed those tokens into your computer every time you want to ask a question.

  • Oka@sopuli.xyz
    link
    fedilink
    arrow-up
    6
    arrow-down
    2
    ·
    2 days ago

    ChatGPT managed to see through my shenanigans:

    If John has six apples and Bob had six oranges, and Bob gave john 2 apples, how many oranges does Betty have if it’s a summer day in january?

    Classic trick question! Let’s break it down:

    John has 6 apples.

    Bob has 6 oranges.

    Bob gives John 2 apples — but wait, Bob didn’t have any apples, only oranges.

    So Bob can’t give John 2 apples.

    Meanwhile, Betty hasn’t even been mentioned until the last sentence, and there’s no info about her oranges.

    Also, “a summer day in January” only makes sense in the southern hemisphere.

    Conclusion: We still have no idea how many oranges Betty has — the question is nonsense on purpose.

    So the answer is either: “Indeterminate”, “Nonsense question”, or “Depends on how much Betty likes oranges in the summer.”

    I think the original message is true for older versions of GPT though, and AI being thrust into everything results in a lot of errors I’ve seen.

    • BudgetBandit@sh.itjust.works
      link
      fedilink
      arrow-up
      8
      ·
      2 days ago

      Also, “a summer day in January” only makes sense in the southern hemisphere.

      According to German news broadcasts, and maybe German meteorologists, a summer day is any day that reaches >25°C. Germany reached a new January record at 18.1°C this year, so another 30 more years and we might get the first summer day of the year in January.

    • callouscomic@lemm.ee
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      2 days ago

      Why can’t Bob give John 2 apples?

      The restriction is merely implied, but we presume Bob did not have anything prior to being given something. Maybe Bob already had them. Bad AI. Lol

      • Shawdow194@fedia.io
        link
        fedilink
        arrow-up
        7
        ·
        1 day ago

        “Depends on how much Betty likes oranges in the summer.”

        It did come up with a quite accurately human and sassy response to the orginal question

  • Boomer Humor Doomergod@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    2
    ·
    edit-2
    2 days ago

    I feel like a really intelligent AI would recognize the context of the text it gets.

    Like, if you or I heard someone saying a bunch of numbers we’d know to switch into math mode and start mathing.

    But these things just dawdle along and start repeating nursery rhymes or whatever because they’re not smart enough to know what they’re doing.

    • Shadow@lemmy.ca
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      2 days ago

      Kinda surprised wolfram alpha hasn’t done more in the ai space

    • pretzelz@lemmy.world
      link
      fedilink
      arrow-up
      3
      arrow-down
      5
      ·
      2 days ago

      Yeah, llms aren’t ai. They are just a fancy Markov model… Need controllers on top to decide when you want to make sentences and when you need to do something else. A controller could be an llm, but a llm by itself is just a tool, not a system

      • LeninsOvaries@lemmy.cafe
        link
        fedilink
        English
        arrow-up
        5
        ·
        2 days ago

        Exactly right. True AI can control an NPC in a video game. Bungie created the perfect AI in 2001. ChatGPT can’t play Halo, so it’s not AI.

  • foremanguy@lemmy.ml
    link
    fedilink
    arrow-up
    4
    arrow-down
    12
    ·
    2 days ago

    I agree with that point of view but at the same time, it’s mostly false. AI resolve problems that cannot be resolved by a dummy calculator. Surely AI are not the best in pure calculations but it’s not there main goal (unless it is explicitly designed for it)

      • foremanguy@lemmy.ml
        link
        fedilink
        arrow-up
        3
        arrow-down
        2
        ·
        2 days ago

        I my opinion you’re partially right. Yeah AI doesn’t solve any problems by itself, humans can already do it. AI is just humans but faster and more efficient (surely this create other serious problems like employment and more…)

        But in this direction is the computer too, resolving no problems, before computers, humans were called computers, and they were doing the job of today iron computers. They didn’t resolve problems too, just faster while creating others important social issues

          • foremanguy@lemmy.ml
            link
            fedilink
            arrow-up
            2
            arrow-down
            2
            ·
            1 day ago

            Humans are not 100% reliable too. Don’t make me wrong, I not promoting the use of AI (on the contrary I’m more against it) But you must admit that I’m totally false in my says. You don’t think that at the time of the first computers people were saying the same thing as today about AI?

  • snooggums@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    5
    ·
    2 days ago

    AI is a shit show because of how it is being terribly implemented and math is the main example.