Jack Riddle[Any/All]

Profile picture drawn by Paws and Claws and licensed under the Creative Commons Attribution Sharealike 4.0 International license(cc-by-sa 4.0)

currently migrating my main account to anise@quokk.au

  • 2 Posts
  • 246 Comments
Joined 10 months ago
cake
Cake day: May 1st, 2025

help-circle
  • what text are you reading that has a 0% error rate?

    as I said, the text has a 0% error rate about the contents of the text, which is what the LLM is summarising, and to which it adds it’s own error rate. Then you read that and add your error rate.

    the question is can we make a system that has an error rate that is close to or lower than a person’s

    can we???

    could you read and summarize 75 novels with a 0% error rate?

    why… would I want that? I read novels because I like reading novels? I also think that on summaries LLMs are especially bad, since there is no distinction between “important” and “unimportant” in the architecture. The point of a summary is to only get the important points, so it clashes.

    provide a page reference to all of the passages written in iambic pentameter?

    no LLM can do this. LLMs are notoriously bad at doing any analysis of this kind of style element because of their architecture. why would you pick this example

    Meanwhile an LLM could produce a summary, with citations generated and tracked by non-AI systems, with an error rate comparable to a human (assuming the human was given a few months to work on the problem) in seconds.

    I still have not seen any evidence for this, and it still does not adress the point that the summary would be pretty much unreadable


  • The study of this in academia

    you are linking to an arxiv preprint. I do not know these researchers. there is nothing that indicates to me that this source is any more credible than a blog post.

    has found that LLM hallucination rate can be dropped to almost nothing

    where? It doesn’t seem to be in this preprint, which is mostly a history of RAG and mentions hallucinations only as a problem affecting certain types of RAG more than other types. It makes some relative claims about accuracy that suggest including irrelevant data might make models more accurate. It doesn’t mention anything about “hallucination rate being dropped to almost nothing”.

    (less than a human)

    you know what has a 0% hallucination rate about the contents of a text? the text

    You can see in the images I posted that it both answered the question and also correctly cited the source which was the entire point of contention.

    this is anecdotal evidence, and also not the only point of contention. Another point was, for example, that ai text is horrible to read. I don’t think RAG(or any other tacked-on tool they’ve been trying for the past few years) fixes that.


  • see, the problem is that I am not going to be reading that text because I know it is unreliable and ai text makes my eyes glaze over, so I will be clicking on all those links until I find something that is reliable. On a search engine I can just click through every link or refine my search with something like site:reddit.com site:wikipedia.org or format:pdf or something similar. With a chatbot, I need to write out the entire question, look at the four or so links it provided and then reprompt it if it doesn’t contain what I’m looking for. I also get a limited amount of searches per day because I am not paying for a chatbot subscription. This is completely pointless to me.
















  • Jack Riddle[Any/All]@lemmy.dbzer0.comto196@lemmy.blahaj.zonegender rule
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    1
    ·
    8 days ago

    just don’t ask for gender? let the player either design a sprite or pick one and pick pronouns or just refer to the player with they/them. Sex is only a fact insofar as the collection of bodily features we commonly refer to as sex exist. the reference to it is still a social choice, and we could’ve just as easily picked another group of features, or none at all.