So I signed up for a free month of their crap because I wanted to test if it solves novel variants of the river crossing puzzle.
Like this one:
You have a duck, a carrot, and a potato. You want to transport them across the river using a boat that can take yourself and up to 2 other items. If the duck is left unsupervised, it will run away.
Unsurprisingly, it does not:
https://g.co/gemini/share/a79dc80c5c6c
https://g.co/gemini/share/59b024d0908b
The only 2 new things seem to be that old variants are no longer novel, and that it is no longer limited to producing incorrect solutions - now it can also incorrectly claim that the solution is impossible.
I think chain of thought / reasoning is a fundamentally dishonest technology. At the end of the day, just like older LLMs it requires that someone solved a similar problem (either online or perhaps in a problem solution pair they generated if they do that to augment the training data).
But it outputs quasi reasoning to pretend that it is actually solving the problem live.
All along my mistake was that I was prompting it in unicode instead of latin1, alphameric BCD, or “modified UTF-8”.
I thought everyone knew that you had to structure prompts in ALGOL 420 to get the best performance by going close to the metal
I use UTF-9 to efficiently handle Unicode on my PDP-10.
@bitofhope @techtakes Surely you need a PDP-9 for that?