Google artificial “intelligence” suggested to add glue to let cheese stick to the pizza, because a decade ago user fucksmith on reddit said so

  • NevermindNoMind@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    6 months ago

    Part of the problem with Google is it’s use of retrieval augmented generation, where it’s not just the llm answering, but the llm is searching for information, apparently through its reddit database from that deal, and serving it as the answer. The tip off is the absurd answers are exact copies of the reddit comments, whereas if the model was just trained on reddit data and responding on its own the model wouldn’t produce verbatim what was in the comments (or shouldn’t, that’s called overfitting and is avoided in the training process). The gemini llm on its own would probably give a better answer.

    The problem here seems to be Google trying to make the answers more trustworthy through rag, but they didn’t bother to scrub the reddit data their relying on well enough, so joke and shit answers are getting mixed in. This is more a datascrubbing problem then an accuracy problem.

    But overall I generally agree with your point.

    One thing I think people overlook though is that for a lot of things, maybe most things, there isn’t a “correct” answer. Expecting llms to reach some arbitrary level of “accuracy” is silly. But what we do need is intelligence and wisdom in these systems. I think the camera jam example is the best illustration of that. Opening the back of the camera and removing the film is technically a correct way to fix the jam, but it ruins the film so it’s not an ideal solution most of the time, but it takes intelligence and wisdom to understand that.