• nightshade [they/them]@hexbear.net
    link
    fedilink
    English
    arrow-up
    10
    ·
    edit-2
    2 months ago

    I’m pretty sure some of the newer ChatGPT-like products (the consumer-facing interface, not the raw LLM) do in fact do this. They try to detect certain types of inputs (i.e. math problems or requesting the current weather) and convert it to an API request to some other service and return the result instead of a LLM output. Frankly it comes across to me as an attempt to make the “AI” seem smarter than it really is by covering up its weaknesses.

    • FunkyStuff [he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 months ago

      Yeah, Siri has been capable of doing that for a long time, but my actual hope would be that moreso than handing the user the API response, the LLM could actually keep operating on that response and do more with it, composing several API calls. But that’s probably prohibitively expensive to train since you’d have to do it billions of times to get the plagiarism machine to learn how to delegate work to an API properly.