Just tried the same prompt and got the same results.
I would assume the way these LLMs are working for search is they're answering based on what they "know" and then backsolving from that to find something to cite to.
Right? I would expect actually performing a search and dumping the results of that search into a prompt and running it through the model would be too resource intensive.