• Re: Trustworthy LLM systems --- require citing external sources ---no more hallucination

    From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Sun Oct 19 16:26:11 2025
    From Newsgroup: comp.theory

    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    They could have reasoning models with mapping into the model from
    context and back out to the context so they have both - a chance that
    you get the reasoning, and how much of the reasoning you get, including something similar to cognitive dissonance.

    A justification of the claim none of them have reasoning would be an interesting read. Automated theorem provers have been around a long time
    and research into making good ones produce proofs we like has been going
    a long time. It's hard to believe after all this capital investment in
    the last few years that they haven't fed statistics from proofs and
    their appeal and conclusions and their effects into theorem provers to
    direct them and from theorem provers to trick you.


    and are required to cite all of their sources

    Such an application exists; for instance Google's NotebookLM.
    It's an application in which you create separate workspaces into which
    you upload documents of your choice.

    Not your free choice, the context is an extension of the model, chatting
    to it is training. Check the copyright on your materials.

    Furthermore, without legal responsibility, or else a locally running
    model with a suitable analysis of its complexity, you can't rely on it.
    Past performance is not an indicator of future performance.


    --
    Tristan Wibberley

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except that you may,
    of course, cite it academically giving credit to me, distribute it
    verbatim as part of a usenet system or its archives, and use it to
    promote my greatness and general superiority without misrepresentation
    of my opinions other than my opinion of my greatness and general
    superiority which you _may_ misrepresent. You definitely MAY NOT train
    any production AI system with it but you may train experimental AI that
    will only be used for evaluation of the AI methods it implements.

    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Sun Oct 19 17:19:58 2025
    From Newsgroup: comp.theory

    On 02/10/2025 00:17, olcott wrote:

    Large language models can do jaw-dropping things. But nobody knows
    exactly why.

    https://www.technologyreview.com/2024/03/04/1089403/large-language-models-amazing-but-nobody-knows-why/

    I don't trust that article. I think the journalist didn't understand and
    mixed referents. If you apply 500% attention you can spot some of the
    mixups but perhaps you have to apply 500% attention to see it.


    --
    Tristan Wibberley

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except that you may,
    of course, cite it academically giving credit to me, distribute it
    verbatim as part of a usenet system or its archives, and use it to
    promote my greatness and general superiority without misrepresentation
    of my opinions other than my opinion of my greatness and general
    superiority which you _may_ misrepresent. You definitely MAY NOT train
    any production AI system with it but you may train experimental AI that
    will only be used for evaluation of the AI methods it implements.

    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Kaz Kylheku@643-408-1753@kylheku.com to comp.theory on Sun Oct 19 18:25:45 2025
    From Newsgroup: comp.theory

    On 2025-10-19, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.
    --
    TXR Programming Language: http://nongnu.org/txr
    Cygnal: Cygwin Native Application Library: http://kylheku.com/cygnal
    Mastodon: @Kazinator@mstdn.ca
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Richard Heathfield@rjh@cpax.org.uk to comp.theory on Sun Oct 19 19:37:10 2025
    From Newsgroup: comp.theory

    On 19/10/2025 19:25, Kaz Kylheku wrote:
    On 2025-10-19, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.

    Trouble is, Kaz, they're getting pretty good at faking it, and
    "pretty good" is good enough for a lot of people. Sometimes you
    have to scratch quite hard to get at the inanity underneath, and
    not everybody wants to scratch; narcissism and gullibility, for
    example, are both contraindications for truth-digging.
    --
    Richard Heathfield
    Email: rjh at cpax dot org dot uk
    "Usenet is a strange place" - dmr 29 July 1999
    Sig line 4 vacant - apply within
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Kaz Kylheku@643-408-1753@kylheku.com to comp.theory on Sun Oct 19 19:51:54 2025
    From Newsgroup: comp.theory

    On 2025-10-19, Richard Heathfield <rjh@cpax.org.uk> wrote:
    On 19/10/2025 19:25, Kaz Kylheku wrote:
    On 2025-10-19, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.

    Trouble is, Kaz, they're getting pretty good at faking it, and
    "pretty good" is good enough for a lot of people. Sometimes you
    have to scratch quite hard to get at the inanity underneath, and
    not everybody wants to scratch; narcissism and gullibility, for
    example, are both contraindications for truth-digging.

    Like I remarked in mhy other posting, when language models are trained,
    the one thing they are learning more than anything from the majority
    of their inputs is *grammar*.

    The inputs are diverse texts about every imaginable thing, but what they
    have in common is grammar in common and so that's what the LLM learns
    best.

    Imagine you're a dummy who got out of high shcool with C's and D's, and
    nobody in your family can write so much as a postcard without grammar
    and spelling errors, right down to people's names. The chatbots must
    seem like geniuses!

    If you're smarter than that, but not versed in a topic, wow,
    the LLM output asbout that topic sure looks smart.

    When you're versed in a topic, wow, shitshow. Not every single time,
    but often enough that it's obvious that LLM is just faking talking
    the talk out of pulling sequences of tokens from training materials.

    In Billy Joel's "Piano Man", the verse goews, "the waitress is
    practicing politics, as the business men slowly get stoned".

    LLMs are exactly like the waitress practicing politics (or any other
    topic she overhears from the regulars at the bar).

    She can interject with something smart sounding in the conversation and,
    man, never mind the piano man, what is /she/ doing here?

    The character of Penny in the sitcom "The Big Bang Theory"
    exploits this trope also. (E.g. Episode 13, Penny helps Sheldon
    solve his equation.)
    --
    TXR Programming Language: http://nongnu.org/txr
    Cygnal: Cygwin Native Application Library: http://kylheku.com/cygnal
    Mastodon: @Kazinator@mstdn.ca
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Chris M. Thomasson@chris.m.thomasson.1@gmail.com to comp.theory on Sun Oct 19 12:58:20 2025
    From Newsgroup: comp.theory

    On 10/19/2025 12:51 PM, Kaz Kylheku wrote:
    On 2025-10-19, Richard Heathfield <rjh@cpax.org.uk> wrote:
    On 19/10/2025 19:25, Kaz Kylheku wrote:
    On 2025-10-19, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.

    Trouble is, Kaz, they're getting pretty good at faking it, and
    "pretty good" is good enough for a lot of people. Sometimes you
    have to scratch quite hard to get at the inanity underneath, and
    not everybody wants to scratch; narcissism and gullibility, for
    example, are both contraindications for truth-digging.

    Like I remarked in mhy other posting, when language models are trained,
    the one thing they are learning more than anything from the majority
    of their inputs is *grammar*.

    The inputs are diverse texts about every imaginable thing, but what they
    have in common is grammar in common and so that's what the LLM learns
    best.

    Imagine you're a dummy who got out of high shcool with C's and D's, and nobody in your family can write so much as a postcard without grammar
    and spelling errors, right down to people's names. The chatbots must
    seem like geniuses!

    If you're smarter than that, but not versed in a topic, wow,
    the LLM output asbout that topic sure looks smart.

    When you're versed in a topic, wow, shitshow. Not every single time,
    but often enough that it's obvious that LLM is just faking talking
    the talk out of pulling sequences of tokens from training materials.

    In Billy Joel's "Piano Man", the verse goews, "the waitress is
    practicing politics, as the business men slowly get stoned".

    LLMs are exactly like the waitress practicing politics (or any other
    topic she overhears from the regulars at the bar).

    How about a clever time share guy...? ;^)



    She can interject with something smart sounding in the conversation and,
    man, never mind the piano man, what is /she/ doing here?

    The character of Penny in the sitcom "The Big Bang Theory"
    exploits this trope also. (E.g. Episode 13, Penny helps Sheldon
    solve his equation.)


    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Mon Oct 20 10:59:59 2025
    From Newsgroup: comp.theory

    On 19/10/2025 20:51, Kaz Kylheku wrote:
    The inputs are diverse texts about every imaginable thing, but what they
    have in common is grammar in common and so that's what the LLM learns
    best.

    Since language interpretation is conventionally just translation to
    other languages I expect semantics to be modelled as language, with its
    grammar (consequences of the laws of physics, consequences of
    conventions and technologies of the built environment, etc).

    What do you think, are semantics special?

    --
    Tristan Wibberley

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except that you may,
    of course, cite it academically giving credit to me, distribute it
    verbatim as part of a usenet system or its archives, and use it to
    promote my greatness and general superiority without misrepresentation
    of my opinions other than my opinion of my greatness and general
    superiority which you _may_ misrepresent. You definitely MAY NOT train
    any production AI system with it but you may train experimental AI that
    will only be used for evaluation of the AI methods it implements.

    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Mon Oct 20 11:22:18 2025
    From Newsgroup: comp.theory

    On 02/10/2025 10:03, David Brown wrote:
    ...someone in a usenet group tells you they have disproved
    a fundamental maths theorem

    I reckon the problem there is popular public/science/math communicators
    and there follows my reckoning of why:

    They deceive the public wrt. the problem statement so the name of the
    problem has two distinct referents. Topical scholars actually receive a disproof of the second, new referent which merely has the same name as
    the first one that said topical scholars know!

    The communicators have learned to deceive because they judge their
    success by the illusion of acceptance by their students who eventually
    accept the mischaracterisations that the tutor experimentally tests. The
    tests are each performed in response to earlier non-acceptance of what
    would be an accurate characterisation but for being given without first establishing terms of the A-language into the U-language of the students!

    The communicator, being insufficiently critical in search of the
    personal heroism of the salvation provided by teaching learns to teach increasingly meme-ishly successful mischaracterisations. They do that by reinforcement learning from the loss of the punishment from the students expressions of failure-to-imbue that the teacher otherwise receives. The teacher goes on to thoroughly screw with every student that follows.

    --
    Tristan Wibberley

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except that you may,
    of course, cite it academically giving credit to me, distribute it
    verbatim as part of a usenet system or its archives, and use it to
    promote my greatness and general superiority without misrepresentation
    of my opinions other than my opinion of my greatness and general
    superiority which you _may_ misrepresent. You definitely MAY NOT train
    any production AI system with it but you may train experimental AI that
    will only be used for evaluation of the AI methods it implements.

    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From olcott@polcott333@gmail.com to comp.theory on Mon Oct 20 10:21:33 2025
    From Newsgroup: comp.theory

    On 10/19/2025 1:37 PM, Richard Heathfield wrote:
    On 19/10/2025 19:25, Kaz Kylheku wrote:
    On 2025-10-19, Tristan Wibberley
    <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.

    Trouble is, Kaz, they're getting pretty good at faking it, and "pretty
    good" is good enough for a lot of people. Sometimes you have to scratch quite hard to get at the inanity underneath, and not everybody wants to scratch; narcissism and gullibility, for example, are both
    contraindications for truth-digging.


    Yes. They got 67-fold smarter in the last two years.
    Two years ago with its 3000 word limit they could
    just barely understand my simplest DD() proof, they
    acted like they has Alzheimer's on my DD() proof.
    They had a 3000 world limit back then. Not it is
    a 200,000 word limit.
    --
    Copyright 2025 Olcott "Talent hits a target no one else can hit; Genius
    hits a target no one else can see." Arthur Schopenhauer
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From olcott@polcott333@gmail.com to comp.theory on Mon Oct 20 10:30:52 2025
    From Newsgroup: comp.theory

    On 10/19/2025 2:51 PM, Kaz Kylheku wrote:
    On 2025-10-19, Richard Heathfield <rjh@cpax.org.uk> wrote:
    On 19/10/2025 19:25, Kaz Kylheku wrote:
    On 2025-10-19, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 01/10/2025 22:12, Kaz Kylheku wrote:
    On 2025-10-01, olcott <polcott333@gmail.com> wrote:

    Not at all.
    When LLM systems have otherwise good reasoning

    They don't have reasoning. They have statistical token prediction,
    plus a few dog-and-pony tricks.

    Did an AI tell you that?

    No.

    Trouble is, Kaz, they're getting pretty good at faking it, and
    "pretty good" is good enough for a lot of people. Sometimes you
    have to scratch quite hard to get at the inanity underneath, and
    not everybody wants to scratch; narcissism and gullibility, for
    example, are both contraindications for truth-digging.

    Like I remarked in mhy other posting, when language models are trained,
    the one thing they are learning more than anything from the majority
    of their inputs is *grammar*.

    The inputs are diverse texts about every imaginable thing, but what they
    have in common is grammar in common and so that's what the LLM learns
    best.

    Imagine you're a dummy who got out of high shcool with C's and D's, and nobody in your family can write so much as a postcard without grammar
    and spelling errors, right down to people's names. The chatbots must
    seem like geniuses!

    If you're smarter than that, but not versed in a topic, wow,
    the LLM output asbout that topic sure looks smart.

    When you're versed in a topic, wow, shitshow. Not every single time,
    but often enough that it's obvious that LLM is just faking talking
    the talk out of pulling sequences of tokens from training materials.

    In Billy Joel's "Piano Man", the verse goews, "the waitress is
    practicing politics, as the business men slowly get stoned".

    LLMs are exactly like the waitress practicing politics (or any other
    topic she overhears from the regulars at the bar).

    She can interject with something smart sounding in the conversation and,
    man, never mind the piano man, what is /she/ doing here?

    The character of Penny in the sitcom "The Big Bang Theory"
    exploits this trope also. (E.g. Episode 13, Penny helps Sheldon
    solve his equation.)


    Clearly you have no recently tested this empirically.

    <Input to LLM systems>
    Please think this all the way through without making any guesses

    Simulating Termination Analyzer HHH correctly simulates its input until:

    (a) It correctly detects that its simulated input cannot
    possibly reach its own simulated final halt state then:
    abort simulation and return 0 rejecting its input as non-halting.

    (b) Simulated input reaches its simulated "return" statement: return 1.

    typedef int (*ptr)();
    int HHH(ptr P);

    int DD()
    {
    int Halt_Status = HHH(DD);
    if (Halt_Status)
    HERE: goto HERE;
    return Halt_Status;
    }

    int main()
    {
    HHH(DD);
    }

    What value should HHH(DD) correctly return?
    </Input to LLM systems>
    --
    Copyright 2025 Olcott "Talent hits a target no one else can hit; Genius
    hits a target no one else can see." Arthur Schopenhauer
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Kaz Kylheku@643-408-1753@kylheku.com to comp.theory on Mon Oct 20 19:18:58 2025
    From Newsgroup: comp.theory

    On 2025-10-20, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    On 19/10/2025 20:51, Kaz Kylheku wrote:
    The inputs are diverse texts about every imaginable thing, but what they
    have in common is grammar in common and so that's what the LLM learns
    best.

    Since language interpretation is conventionally just translation to
    other languages I expect semantics to be modelled as language, with its grammar (consequences of the laws of physics, consequences of
    conventions and technologies of the built environment, etc).

    What do you think, are semantics special?

    OK, so if you have some equations whose solution indicates the paths of
    three bodies in space, those paths are not what the equations are
    actually about (i.e. their semantics). Their semantics is always just a translation into some string of symbols in another language.

    P.S. "semantics" is singular, like mathematics, politics, ...
    --
    TXR Programming Language: http://nongnu.org/txr
    Cygnal: Cygwin Native Application Library: http://kylheku.com/cygnal
    Mastodon: @Kazinator@mstdn.ca
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Tue Oct 21 09:19:29 2025
    From Newsgroup: comp.theory

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except as noted in
    the sig.

    On 20/10/2025 20:18, Kaz Kylheku wrote:
    ... so if you have some equations whose solution indicates the paths of
    three bodies in space, those paths are not what the equations are
    actually about.

    Wellll.. "being about [something]" at least isn't intrinsic to the
    equations themselves, is it, but some world in which the interpreting
    brain is embodied? Perhaps that just makes the compositional theory of
    meaning of words (using lambda expressions) rather a trivial idea - a
    parameter is needed.

    --
    Tristan Wibberley

    The message body is Copyright (C) 2025 Tristan Wibberley except
    citations and quotations noted. All Rights Reserved except that you may,
    of course, cite it academically giving credit to me, distribute it
    verbatim as part of a usenet system or its archives, and use it to
    promote my greatness and general superiority without misrepresentation
    of my opinions other than my opinion of my greatness and general
    superiority which you _may_ misrepresent. You definitely MAY NOT train
    any production AI system with it but you may train experimental AI that
    will only be used for evaluation of the AI methods it implements.

    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Tristan Wibberley@tristan.wibberley+netnews2@alumni.manchester.ac.uk to comp.theory on Tue Oct 21 09:20:50 2025
    From Newsgroup: comp.theory

    Kaz, I note we're getting quite off-topic.

    --
    Tristan Wibberley
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Kaz Kylheku@643-408-1753@kylheku.com to comp.theory on Tue Oct 21 16:12:46 2025
    From Newsgroup: comp.theory

    On 2025-10-21, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    Kaz, I note we're getting quite off-topic.

    Right, sorry. Where were we, again?

    HH(DD) wrongly returns 0, because DD halts.
    --
    TXR Programming Language: http://nongnu.org/txr
    Cygnal: Cygwin Native Application Library: http://kylheku.com/cygnal
    Mastodon: @Kazinator@mstdn.ca
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From Richard Heathfield@rjh@cpax.org.uk to comp.theory on Tue Oct 21 17:34:05 2025
    From Newsgroup: comp.theory

    On 21/10/2025 17:12, Kaz Kylheku wrote:
    On 2025-10-21, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    Kaz, I note we're getting quite off-topic.

    Right, sorry. Where were we, again?

    HH(DD) wrongly returns 0, because DD halts.


    It should, of course, return 1.

    Oh, wait...
    --
    Richard Heathfield
    Email: rjh at cpax dot org dot uk
    "Usenet is a strange place" - dmr 29 July 1999
    Sig line 4 vacant - apply within
    --- Synchronet 3.21a-Linux NewsLink 1.2
  • From olcott@polcott333@gmail.com to comp.theory on Tue Oct 21 20:46:55 2025
    From Newsgroup: comp.theory

    On 10/21/2025 11:12 AM, Kaz Kylheku wrote:
    On 2025-10-21, Tristan Wibberley <tristan.wibberley+netnews2@alumni.manchester.ac.uk> wrote:
    Kaz, I note we're getting quite off-topic.

    Right, sorry. Where were we, again?

    HH(DD) wrongly returns 0, because DD halts.


    <Input to LLM systems>
    Please think this all the way through without making any guesses.
    Only report on the behavior observed in simulation.

    Simulating Termination Analyzer HHH correctly simulates its input until:
    (a) Detects a non-terminating behavior pattern:
    abort simulation and return 0.
    (b) Simulated input reaches its simulated "return" statement:
    return 1.

    typedef int (*ptr)();
    int HHH(ptr P);

    int DD()
    {
    int Halt_Status = HHH(DD);
    if (Halt_Status)
    HERE: goto HERE;
    return Halt_Status;
    }

    int main()
    {
    HHH(DD);
    }

    What value should HHH(DD) correctly return?
    </Input to LLM systems>
    --
    Copyright 2025 Olcott "Talent hits a target no one else can hit; Genius
    hits a target no one else can see." Arthur Schopenhauer
    --- Synchronet 3.21a-Linux NewsLink 1.2