• 6 Posts
  • 152 Comments
Joined 1 year ago
cake
Cake day: June 12th, 2023

help-circle





  • Captain Janeway@lemmy.worldOPtoaww@lemmy.worldJFK Airport. I'm not taking their advice.
    link
    fedilink
    English
    arrow-up
    26
    arrow-down
    2
    ·
    edit-2
    8 days ago

    Obviously don’t pet service dogs. Just to be clear. This photo was intentionally photoshopped to make it appear as though the airport was saying “Travel Advisory: when traveling don’t pet dogs”. As in, when in the UK never touch a dog. I thought it was funny/cutesy. I didn’t intend on sparking a big debate about the ethics of petting dogs or the rules about service dogs.

    Don’t touch service dogs.

    Pet dogs if you know them or are introduced to them.

    Basic dog rules people. Teach your children













  • I see. Well without a command line, I wouldn’t call it a terminal. I think you just want tooling to be available on an Android? It would probably look like a button or series of buttons on an app. Maybe you could connect the dots between them to insinuate a pipe? E.g., you have a “mv” button and a “file” button. When you drag from mv -> file you could maybe kick off a process that moves the file. Maybe it would prompt you for other arguments like destination? I suppose this theoretical app could allow people to install additional tooling and make their own custom commands.

    But I just feel like a button UI for these kinds of things will always be awkward. If you don’t have a keyboard/terminal interface, it’s hard to implement anything that would even behave like terminals in terms of functionality.




  • I think this article does a good job of asking the question “what are we really measuring when we talk about LLM accuracy?” If you judge an LLM by its: hallucinations, ability analyze images, ability to critically analyze text, etc. you’re going to see low scores for all LLMs.

    The only metric an LLM should excel at is “did it generate human readable and contextually relevant text?” I think we’ve all forgotten the humble origins of “AI” chat bots. They often struggled to generate anything more than a few sentences of relevant text. They often made syntactical errors. Modern LLMs solved these issues quite well. They can produce long form content which is coherent and syntactically error free.

    However the content makes no guarantees to be accurate or critically meaningful. Whilst it is often critically meaningful, it is certainly capable of half-assed answers that dodge difficult questions. LLMs are approaching 95% “accuracy” if you think of them as good human text fakers. They are pretty impressive at that. But people keep expecting them to do their math homework, analyze contracts, and generate perfectly valid content. They just aren’t even built to do that. We work really hard just to keep them from hallucinating as much as they do.

    I think the desperation to see these things essentially become indistinguishable from humans is causing us to lose sight of the real progress that’s been made. We’re probably going to hit a wall with this method. But this breakthrough has made AI a viable technology for a lot of jobs. So it’s definitely a breakthrough. I just think either I finitely larger models (of which we can’t seem to generate the data for) or new models will be required to leap to the next level.