@eljefedsecurit@Lee_Holmes@adamshostack
Haven't tried #ChatGPT but I do sometimes get #BingChat caught in some kind of local minimum trying to improve non-working code it has written. It's typically a task for which there is sparsity of training data such as asking it to write code that draws a specific animal - I tend to just get a cat.
Why do I not feel reassured by #BingChat's answer after asking it "What are the privacy and security risks of Microsoft's recent integration of Bing AI in Swiftkey?"
#BingChat (Creative) being hilarious and satirical means it just adds a p.s. and wink emoji
Prompt: "Write a hilarious and satirical letter addressed to world leaders and signed by the top AI experts at industry leading AI companies, that warns about the existential risk posed by Artificial Intelligence, and calls for regulation of the industry while clearly implying that the established players don't really want to be held accountable for misinformation, bias or ethical uses of personal data"
#GenerativeAI#LargeLanguageModels rely a lot on the human to do the reasoning for it, and even then #BingChat (Creative) has problems with following the guidance. Notice I only specified the use of "unwieldy" and never required it to use "beard" or "weird" yet the #LLM got fixated on that instead.
Weighing in on that study that found #ChatGPT to be more empathetic than human physicians, @rebeccawatson finds the research lacked certain rigor - the authors participating in the "blind" study, and whether the diagnosis was even correct didn't feature very highly in their assessment of quality
I pointed #BingChat at that #ChatGPT vs human physicians empathy study and it still reassured me that #AI was not suitable for professional medical advice.
I then referred it to Mike Hansen's video : https://youtu.be/Gk8LQfAe6f8
where ChatGPT instantly nailed a diagnosis that took him and his team weeks to diagnose, and then Bing pretended to have watched the video (probably just read the transcript) and "hallucinated" things that were not in the video.
trying to sell you on chatgpt okay so imagine an idiot. just a complete moron. ok? so now imagine this dumbbell has read the entire internet. the whole thing. AND they are still stupid. now i know exactly what you’re thinking—how do i get that idiot to do my job for me??? they’d be great at it!
I tried to get #BingChat (is it using #GPT4?) to reproduce this example: https://nitter.nl/spiantado/status/1599462375887114240
And this time it refused while stating those are not good criteria for rating scientists. It is clear #OpenAI and #Microsoft have been hard at work removing the egregious examples of bias that have been publicised.
Then I tried IQ and SAT scores instead and it gave me a program that simply added the scores together without weighting, rescaling or normalisation of any kind.