wick, 2 months ago I guess I just didn’t know that LLMs were set up his way. I figured they were fed massive hash tables of behaviour directly into their robot brains before a text prompt was even plugged in. But yea, tested it myself and got the same result.
I guess I just didn’t know that LLMs were set up his way. I figured they were fed massive hash tables of behaviour directly into their robot brains before a text prompt was even plugged in.
But yea, tested it myself and got the same result.