• 1 Post
  • 99 Comments
Joined 2 years ago
cake
Cake day: June 3rd, 2023

help-circle










  • AndrasKrigare@beehaw.orgto196@lemmy.blahaj.zonefisherman rule
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    3 months ago

    Definitely a fair question; at least for the large bodies of water near me, before they let you park and take your vessel in, they check the registration on it. I’m doing that, they can check what other bodies of water you’ve taken the vessel to and how recently, and can then determine whether or not it’s safe for you to bring it into that one.

    I don’t actually know what the recourse is if it’s not safe, if they make you spray it down thoroughly there, or completely refuse entry, or something else. My information here is from a vague memory of a sign at a lake parking area that was titled “Why do I need to register?” or something like that.









  • I think to some extent it’s a matter of scale, though. If I advertise something as a calculator capable of doing all math, and it can only do one problem, it is so drastically far away from its intended purpose that the meaning kinda breaks down. I don’t think it would be wrong to say “it malfunctions in 99.999999% of use cases” but it would be easier to say that it just doesn’t work.

    Continuing (and torturing) that analogy, if we did the disgusting work of precomputing all 2 number math problems for integers from -1,000,000 to 1,000,000 and I think you could say you had a (really shitty and slow) calculator, which “malfunctions” for numbers outside that range if you don’t specify the limitation ahead of time. Not crazy different from software which has issues with max_int or small buffers.

    If it were the case that there had only been one case of a hallucination with LLMs, I think we could pretty safely call that a malfunction (and we wouldn’t be having this conversation). If it happens 0.000001% of the time, I think we could still call it a malfunction and that it performs better than a lot of software. 99.999% of the time, it’d be better to say that it just doesn’t work. I don’t think there is, or even needs to be, some unified understanding of where the line is between them.

    Really my point is there are enough things to criticize about LLMs and people’s use of them, this seems like a really silly one to try and push.