reedmideke,
@reedmideke@mastodon.social avatar

On that CNET thing in the last boost, my first thought was "this is gonna make search even more useless" and… yeeeep "They are clearly optimized to take advantage of Google’s search algorithms, and to end up at the top of peoples’ results pages"

https://gizmodo.com/cnet-ai-chatgpt-news-robot-1849996151

reedmideke,
@reedmideke@mastodon.social avatar

I guess on the plus side, it'll encourage search engines to put resources into detecting AI generated garbage? Because even if CNET is eventually shamed into quitting, blackhat SEOers are gonna be all over it

reedmideke,
@reedmideke@mastodon.social avatar

Man this just gets wilder. If accurate, CNET pretty much already went full blackhat SEO https://futurism.com/cnet-ai-articles-label

reedmideke,
@reedmideke@mastodon.social avatar

Surprise! The dudes turning CNET into a low rent SEO farm weren't too picky about the quality of their "AI" https://futurism.com/cnet-ai-plagiarism

reedmideke,
@reedmideke@mastodon.social avatar

Kicker in the latest installment in @jonchristian's coverage of the CNET AI saga… private equity bro's solution to the AI scandal is to rebrand their "AI" as "Tooling" (we also get confirmation said "tooling" comes from OpenAI)

https://futurism.com/leaked-messages-cnet-red-ventures-ai

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Nice @cfiesler article on prompt hacking. Kinda boggles my mind that in the year 2023, a bunch of the biggest names in tech decided that free-form, in-band commands were an acceptable way to control these things.

Yes I realize the nature of the tech makes it very hard to do otherwise, but still… did y'all sleep through the last 20 years of SQL injection and XSS?

https://kotaku.com/chatgpt-ai-openai-dan-censorship-chatbot-reddit-1850088408

reedmideke,
@reedmideke@mastodon.social avatar

Not like I'd trust human written medical advice in these rags, but of all the topic to hand over to a bullshit machine, this seems like an especially poor one
https://futurism.com/neoscope/magazine-mens-journal-errors-ai-health-article

reedmideke,
@reedmideke@mastodon.social avatar

Post from @opencage is a good illustration of how these things are (Frankfurtian sense) bullshit machines. If users asked about the goecoding service opencage actually offers, there's a fair chance they could get a reasonable answer. Ask about an adjacent, but non-existent service, and ChatGPT makes one up based on similar things in the training set https://blog.opencagedata.com/post/dont-believe-chatgpt

reedmideke,
@reedmideke@mastodon.social avatar

LLMs are genuinely amazing, but a lot of people hyping them seem assume this kind of thing is just a bug that needs to be fixed ("We're 90% of the way there, just have to clean up these last few issues!") rather than an inherent characteristic of the technology… optimism which does not seem to be justified on any technical or theoretical grounds

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

This is wild. There's been a lot of overblown hype about the threat of AI in disinfo but this is one I expected: Flooding the zone with superficially plausible looking sources / backstory, so human created content will look well sourced to people who don't check. Yet somehow even stupider and lazier than I imagined. Anyway, reminder The Grayzone is trash https://twitter.com/AricToler/status/1635666413510701057

reedmideke,
@reedmideke@mastodon.social avatar

"Microsoft announced Tuesday that the Bing AI chatbot, released last month, had been using GPT-4 all along"- wait, the new AI hype thing is actually the old AI hype thing everyone was dunking on last week? https://www.washingtonpost.com/technology/2023/03/14/gpt-4-has-arrived-it-will-blow-chatgpt-out-water/

reedmideke,
@reedmideke@mastodon.social avatar

Also, I really wish @drewharwell would have pushed OpenAI to describe more specifically what they mean by "advanced reasoning capabilities" … how do they define "reasoning"? Is there some objective or theoretical basis to assert LLMs can "reason" in the conventional sense of the word?

reedmideke,
@reedmideke@mastodon.social avatar

admits "still makes many of the errors of previous versions, including 'hallucinating' nonsense, perpetuating social biases and offering bad advice" but implies (largely unchallenged in the press) that these are merely bugs to be fixed, not fundamental characteristics of the technology

reedmideke,
@reedmideke@mastodon.social avatar

"we translated the MMLU benchmark—a suite of 14,000 multiple-choice problems spanning 57 subjects—into a variety of languages using Azure Translate"
So they tested multi-lingual capability… using ML based machine translations? All those billions in venture capital, and they're to f-ing cheap to hire human translators? https://openai.com/research/gpt-4

reedmideke,
@reedmideke@mastodon.social avatar

I mean, maybe all 14k problems in every language is too much, but at least use a human validated subset?

reedmideke,
@reedmideke@mastodon.social avatar

So @Riedl found you can trivially influence bing chat with hidden text

Thought one: Whoohoo, we've re-invented early 2000s keyword stuffing

https://mastodon.social/@Riedl@sigmoid.social/110058596766522240

reedmideke,
@reedmideke@mastodon.social avatar

Thought two: If you could convince bing to produce arbitrary output in a response, what's the worst you can do? Just some Bing-specific text? XSS is presumably out (barring bugs, the app has to be prepared for arbitrary output anyway). Can you make it generate a url that leaks some information about the chatter?

reedmideke,
@reedmideke@mastodon.social avatar

Anyway, bing chat apparently doesn't have a good memory, but will cheerfully lie about it

reedmideke,
@reedmideke@mastodon.social avatar

This from @cfarivar on the other site is also a great illustration of how these things confidently make stuff up https://twitter.com/cfarivar/status/1638211525449248770

reedmideke,
@reedmideke@mastodon.social avatar

I remain baffled as to why anyone would want their search results summarized in conversational language by an extremely confident bullshitter ¯_(ツ)_/¯

reedmideke,
@reedmideke@mastodon.social avatar

One argument is "they're getting better so eventually they'll be good enough" but, without some fundamental grounding in fact, does reducing the amount of obvious bullshit make them reliable, or just train them to produce more subtle bullshit, or only bullshit on more obscure topics?

reedmideke,
@reedmideke@mastodon.social avatar

Like, saying Novato and Santa Monica are 80 miles apart is a big obvious error, but is a model that gets that right, but makes similar errors about small, obscure towns in the midwest "better" ?

reedmideke,
@reedmideke@mastodon.social avatar

And again, who needs this (previous boost, lack of QT really really ain't great for running threads like this)? Also, while the other answers seem roughly correct, how sure would you be without googling? And if you have to google each one, what's the point? https://mastodon.social/@caseynewton/110063518076189757

reedmideke,
@reedmideke@mastodon.social avatar

Turnitin claims their AI text detector has a 1% false positive rate… meaning a typical high school English class relying on it could have hundreds of false accusations of misconduct in a semester. They spin that as good enough for market https://www.washingtonpost.com/technology/2023/04/01/chatgpt-cheating-detection-turnitin/

reedmideke,
@reedmideke@mastodon.social avatar

"Turnitin also says its scores should be treated as an indication, not an accusation"

  1. Good luck communicating that nuance to thousands of teachers
  2. Creates an major opportunity to amplify bias
reedmideke,
@reedmideke@mastodon.social avatar

Who could have predicted this? "But producing images of a cartoon character like Mickey Mouse—even wholly original images—may infringe Disney’s copyrights" https://arstechnica.com/tech-policy/2023/04/stable-diffusion-copyright-lawsuits-could-be-a-legal-earthquake-for-ai/

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

"Wallace pointed out that Stable Diffusion is only a few gigabytes in size—far too small to contain compressed copies of all or even very many of its training images"

Have to say I'm not very convinced by this "it doesn't store the images" argument. If it can produce an image that would considered infringing in other contexts, whether you can point to that image in the blob of ML data seems rather beside the point

reedmideke,
@reedmideke@mastodon.social avatar

.@willoremus digs into the actual near term threat (spoiler, it's not skynet)

"761 [digital marketing firm clients] said they’ve at least experimented with some form of generative AI to produce online content, while 370 said they now use it to help generate most if not all of their new content"

https://wapo.st/42qYNW7 (gift link)

reedmideke,
@reedmideke@mastodon.social avatar

"Ingenio, the San Francisco-based online publisher behind sites such as horoscope dot com and astrology dot com, is among those embracing automated content"
Hmm, are AI generated horoscopes more BS than the real thing? 🤔

reedmideke,
@reedmideke@mastodon.social avatar

But again, the real issue is volume: "We published a celebrity profile a month. Now we can do 10,000 a month."

reedmideke,
@reedmideke@mastodon.social avatar

"One sees Alexa creating a bedtime story using a prompt from a kid…"
Oh yeah, don't see ANY way having LLMs come up with bedtime stories for your kids could possibly go wrong https://arstechnica.com/gadgets/2023/05/as-alexa-flounders-amazon-hopes-homegrown-generative-ai-can-find-it-revenue/

reedmideke,
@reedmideke@mastodon.social avatar

This from @waldoj is a really excellent example of how these things BS https://mastodon.social/@waldoj/110353407663057558

reedmideke,
@reedmideke@mastodon.social avatar

I've seen people described LLMs as "recognizing" or "admitting" they were wrong when pressed on a BS answer, but of course, that's just because admitting a mistake is one probable response to having an error pointed out.

They are likely tweaked against the alternative of continuing to argue, because being aggressively wrong is a bad look (except that one asshole version of bing everyone mocked)

reedmideke,
@reedmideke@mastodon.social avatar

Ironically, asshole bing is probably more representative of a training set derived from internet text, so politely accepting your correction is presumably a result of deliberate effort https://www.voanews.com/a/angry-bing-chatbot-just-mimicking-humans-experts-say-/6969343.html

reedmideke,
@reedmideke@mastodon.social avatar

Who could have seen this coming? Turns out asking a stochastic bullshit machine whether it wrote a thing is not an accurate way to determine whether it actually wrote the thing (gift link) https://wapo.st/45eXjAl

reedmideke,
@reedmideke@mastodon.social avatar

Oh my. A lawyer used output in their filings and it's going about as well as you'd expect (presuming you have a couple brain cells to rub together)

https://twitter.com/steve_vladeck/status/1662286888890138624

(filings https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/)

reedmideke,
@reedmideke@mastodon.social avatar

"A submission filed by plaintiff’s counsel in opposition to a motion to dismiss is replete with citations to non-existent cases… the Court issued Orders requiring plaintiff’s counsel to provide an affidavit annexing copies of certain judicial opinions of courts of record cited in his submission, and he has complied… Six of the submitted submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations"
https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-31

reedmideke,
@reedmideke@mastodon.social avatar

So not only did he use to write the original filing, when called on the bogus citations he used ChatGPT to generate the supposed decisions in the cited (non-existent) cases 🤯 (they're in https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-29)

reedmideke,
@reedmideke@mastodon.social avatar

On the one hand, I have trouble believing Schwartz' "I had no idea would make shit up" defense, but on the other, did he really think opposing counsel wouldn't notice, after they already called him on the bogus citations?

reedmideke,
@reedmideke@mastodon.social avatar

The original "hey we couldn't find any of those cases" was in entry https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-24

In which we also learn what the case is about: "There is no dispute that the Plaintiff was travelling as a passenger on an international flight when he allegedly sustained injury after a metal serving cart struck his left knee"

… two dudes set their law licenses on fire for a personal injury suit for a guy who took a drink cart to the knee?

reedmideke,
@reedmideke@mastodon.social avatar

Idle thoughts: In a legal context, this sort of stuff is likely to be caught pretty quickly.

As happened here, the opposing side is going to try to find the cited cases and notice if they're like, totally made up.

So aside from the poor plaintiff who hired these clowns (and presumably has an argument for inadequate representation), the risk should be limited… but a lot of other contexts are much less well positioned to catch plausible looking BS early

reedmideke,
@reedmideke@mastodon.social avatar

Also, while "Varghese v. China Southern Airlines" and friends are unlikely to slip into authoritative sources as a real cases, it wouldn't be at all surprising for general search engines or future LLMs to pick it up and fail to recognize it isn't real

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Very good play-by-play on the lawyers from @kendraserra (this is where I wish we had proper quote toots) https://mastodon.social/@kendraserra@dair-community.social/110441210421818852

reedmideke,
@reedmideke@mastodon.social avatar

Finally, an article that at least raises the question whether BSing may be an inherent characteristic of LLMs rather than a bug that can be fixed (gift link)

https://wapo.st/43obDFt

reedmideke,
@reedmideke@mastodon.social avatar

The "solutions" discussed mostly strike me as bandaids: "a system they called “SelfCheckGPT” that would ask the same bot a question multiple times, then tell it to compare the different answers. If the answers were consistent, it was likely the facts were correct"

reedmideke,
@reedmideke@mastodon.social avatar

and "researchers proposed using different chatbots to produce multiple answers to the same question and then letting them debate each other until one answer won out"

reedmideke,
@reedmideke@mastodon.social avatar

Seems like these might reduce glaring errors where the training data contains a clear consensus correct answer, but doesn't really address the underlying problem.

Is a model that's usually right about stuff "everyone knows" while still making shit up about less obvious topics an improvement? Or does being right about obvious stuff encourage people to trust it when the shouldn't?

reedmideke,
@reedmideke@mastodon.social avatar

Meanwhile that "Air force AI attacks operator in simulation" story is entertaining, but hardly seems representative of any potential real world usage or risks https://www.vice.com/en/article/4a33gj/ai-controlled-drone-goes-rogue-kills-human-operator-in-usaf-simulated-test

reedmideke,
@reedmideke@mastodon.social avatar

lol. 'the "rogue AI drone simulation" was a hypothetical "thought experiment" from outside the military'

https://arstechnica.com/information-technology/2023/06/air-force-denies-running-simulation-where-ai-drone-killed-its-operator/

reedmideke,
@reedmideke@mastodon.social avatar

So, people cosplaying as killer behaved like stereotypical sci-fi killer AI, clearly demonstrating the existential threat of killer AI!

reedmideke,
@reedmideke@mastodon.social avatar

Seriously that @davidgerard piece has it all, but I liked this illustration of bollockschain to AI pipeline "IBM: “The convergence of AI and blockchain brings new value to business.” IBM previously folded its failed blockchain unit into the unit for its failed Watson AI"

https://davidgerard.co.uk/blockchain/2023/06/03/crypto-collapse-get-in-loser-were-pivoting-to-ai/

reedmideke,
@reedmideke@mastodon.social avatar

's lawyers have filed their response, arguing that while their clients may have been extremely reckless and incompetent, they did not know the cases were fake, and so don't meet the "subjective bad faith" standard for sanctions https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-45

reedmideke,
@reedmideke@mastodon.social avatar

and TBH, I kinda believe them, because as stupid as they were, knowingly trying to pass off completely fake cases would seem even stupider. Still mindboggling you could get that far and not check though

reedmideke,
@reedmideke@mastodon.social avatar

@willoremus digs into the compute cost of and boy does that not look like good news for all the startups cramming into everything (gift link) https://wapo.st/3WTCK8Q

reedmideke,
@reedmideke@mastodon.social avatar

Also, pity the poor gamers who have only recently started to see GPU availability recover from the cryptominer induced shortages

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

This blow by blow over on the bird site suggests he took an extremely dim view of 's buddy LoDuca who was signing off on the filings without reading them. Also sounds like they fibbed about who was on vacation when they asked for the extension 😬

https://twitter.com/innercitypress/status/1666838526762139650

reedmideke,
@reedmideke@mastodon.social avatar

Filing what appears to be a thinly veiled pitch for a law-oriented AI startup as an amicus on this case is… a choice https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-50

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Low wage clickworkers surreptitiously using to do their tasks

reedmideke,
@reedmideke@mastodon.social avatar

ruling is in and it's… SANCTIONS FOR EVERYONE! Unsurprisingly, the judge didn't buy the "no bad faith" argument, for predictable reasons such as
"Above Mr. LoDuca’s signature line, the Affirmation in Opposition states, “I declare under penalty of perjury that the foregoing is true and correct”
Although Mr. LoDuca signed the Affirmation in Opposition and filed it on ECF, he was not its author"

https://www.courtlistener.com/docket/63107798/mata-v-avianca-inc/#entry-54

reedmideke,
@reedmideke@mastodon.social avatar

Also, not great plan to lie about being on vacation when responding to a show cause order "Mr. LoDuca’s statement was false and he knew it to be false at the time he made the statement. Under questioning by the Court at the sanctions hearing, Mr. LoDuca admitted that he was not out of the office on vacation"

reedmideke,
@reedmideke@mastodon.social avatar

Mr. Schwartz fares no better
"Mr. Schwartz’s statement in his May 25 affidavit that ChatGPT “supplemented” his research was a misleading attempt to mitigate his actions by creating the false impression that he had done other, meaningful research on the issue and did not rely exclusive on an AI chatbot, when, in truth and in fact, it was the only source"

reedmideke,
@reedmideke@mastodon.social avatar

More on clickworkers allegedly using to automate their AI training tasks. Back door Habsburg AI https://www.technologyreview.com/2023/06/22/1075405/the-people-paid-to-train-ai-are-outsourcing-their-work-to-ai/

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Janelle Shane on the recent detector paper, with succinct advice "Don't use AI detectors for anything important"

https://www.aiweirdness.com/dont-use-ai-detectors-for-anything-important/

reedmideke,
@reedmideke@mastodon.social avatar

This. This is what people reporting on / hype need to understand
https://mastodon.social/@amydentata@tech.lgbt/110651829564300496

reedmideke,
@reedmideke@mastodon.social avatar

Above could also have helped @mozilla avoid the whole explain train wreck, which thankfully seems to have been rolled back https://github.com/mdn/yari/issues/9208#issuecomment-1615411943

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Additional comment from io9 deputy editor James Whitbrook "that's the formal part, here's my own personal comment: lmao, it's fucking dogshit"
https://twitter.com/Jwhitbrook/status/1676704102754004996

reedmideke,
@reedmideke@mastodon.social avatar

Oh FFS @mozilla @stevetex "readers also pointed out a handful of concrete cases where an incorrect answer was rendered. This feedback is enormously helpful, and the MDN team is now investigating these bug reports"

They aren't "bugs" - by definition just put together plausible sounding words with no regard to correctness. Pointing out individual errors demonstrates this, but does not provide any mechanism by which it might be "fixed" in the general case

https://blog.mozilla.org/en/products/mdn/responsibly-empowering-developers-with-ai-on-mdn/

reedmideke,
@reedmideke@mastodon.social avatar

The post also notes that many users were happy with the answers, ignoring that the target audience of people who came to MDN looking for help with something they didn't already know may not immediately recognize that the answer is subtly wrong, or just plausible looking gibberish

reedmideke,
@reedmideke@mastodon.social avatar

It also says "even extraordinarily well-trained LLMs — like humans — will sometimes be wrong"

which is true as far as it goes, but here's the thing: They are not wrong like humans … yes, you'll find some overconfident bullshitters on stack overflow, but generally humans in these contexts have some awareness of the limits of their knowledge and don't drift seamlessly between accurate explanation and complete BS

reedmideke,
@reedmideke@mastodon.social avatar

@stevetex post also makes no mention of the apparent lack of communication with the rest of the MDN team https://github.com/mdn/yari/issues/9208#issuecomment-1615411943

reedmideke,
@reedmideke@mastodon.social avatar

Anyway, there's a new bug, so if you have thoughts on adding stochastic bullshit to what has, up to now, been the premier technical reference for web developers, you could make them heard there https://github.com/mdn/yari/issues/9230

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

More on the debacle: After publishing error-ridden garbage which their own editorial team called "fucking dogshit" 'a G/O Media spokesman, said the company would be “derelict” if it did not experiment with AI. “We think the AI trial has been successful,”'

(free link)

https://wapo.st/43iHlmP

reedmideke,
@reedmideke@mastodon.social avatar

is going great
(caveat I don't know the source and thought it might be a joke, but the rest of their timeline looks real, and Janelle Shane retweeted it)
https://twitter.com/guntrip/status/1640694869785030657

reedmideke,
@reedmideke@mastodon.social avatar

Not gonna screenshot the thread of screenshots here, but it's archived if you don't want to visit the bird site https://web.archive.org/web/20230329232724/https://twitter.com/guntrip/status/1640694869785030657

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

A thing that occurs to me about that last boost from @zoe (https://mastodon.social/@zoe@social.animeprincess.net/110797643482092764): scrapers refusing to play nice with ROBOTS.TXT is that it encourages adversarial approaches…

reedmideke,
@reedmideke@mastodon.social avatar

People building models will be keen to exclude AI generated content from the training set. So, would interspersing stuff that scores high as AI-generated (whether it actually is or not) cause entire pages to be excluded? You could separate it from the real content in ways that humans would understand. OTOH, if you care about SEO it'd be pretty risky

reedmideke,
@reedmideke@mastodon.social avatar

There's also been talk about standards to identify AI generated content, leading to hilarious option of falsely identifying your real content as AI generated to stop people from training AI on it

reedmideke,
@reedmideke@mastodon.social avatar

Folks have suggested CSS based approaches to poison models (like white text on white background) but there's a significant risk of breaking accessibility. Also risk of search engines thinking it looks spammy again

reedmideke,
@reedmideke@mastodon.social avatar

A general problem with poisoning like this is that any technique which becomes really widespread will likely be noticed and filtered out. OTOH, if the goal is to not have your content used, that may be OK!

reedmideke,
@reedmideke@mastodon.social avatar

Good to see mainstream press finally touching the question of whether BSing is fixable or an inherent property of the tech, even if it gets a bit of he said, she said treatment.

Also uh "Those errors are not a huge problem for the marketing firms turning to Jasper AI for help writing pitches…" marketing doesn't care if their pitches are BS? KNOCK ME OVER WITH A FEATHER

https://fortune.com/2023/08/01/can-ai-chatgpt-hallucinations-be-fixed-experts-doubt-altman-openai/

reedmideke,
@reedmideke@mastodon.social avatar

So @Toke points out (https://mastodon.social/@Toke@helvede.net/110848880977610283) that does claim to have unique user agent and honor robots.txt when scraping text for training. Not clear whether this is the only or even primary way publicly accessible web content gets into their training set though https://platform.openai.com/docs/gptbot

reedmideke,
@reedmideke@mastodon.social avatar

Just hypothetically speaking, many web platforms could easily be configured to serve specially tailored content based on the user agent, but that would be mean and wrong and potentially waste resources of VC backed billionaires freeloading off the public web to build their BS machines so definitely don't do that 😉

reedmideke,
@reedmideke@mastodon.social avatar

Complete gibberish will likely get weeded out. Common knowledge will tend to be overwhelmed by other sources. So the sweet spot for influence would seem to be obscure topics, or unique tokens that only appear in your content (though to what end isn't obvious).

Bring on the SolidGoldMagikarp https://www.lesswrong.com/posts/Ya9LzwEbfaAMY8ABo/solidgoldmagikarp-ii-technical-details-and-more-recent

reedmideke,
@reedmideke@mastodon.social avatar

Of course, these things don't just scrape human readable text, many of them do code too. Serving up a special vulnerable version of your input sanitization code when you see GPTBot is left as an exercise to the reader

reedmideke,
@reedmideke@mastodon.social avatar

"It's highly unlikely that ChatGPT's training data includes the entire text of each book under question, though the data may include references to discussions about the book's content—if the book is famous enough"
Highlights a pernicious problem with ChatGPT style : It's far more likely to give reasonable answers on well-known subjects. If you spot check with say, Dickens and Hunter S. Thomson, you might think it was pretty good at spotting naughty books

https://arstechnica.com/information-technology/2023/08/an-iowa-school-district-is-using-chatgpt-to-decide-which-books-to-ban/

reedmideke,
@reedmideke@mastodon.social avatar

But for more obscure ones, it's probably no better than a coin toss. Being relatively good at stuff "everyone knows" gives people false confidence that it's also good at stuff they don't know

(we should also note that even if the entire text of the books were in the training set, that wouldn't mean it would provide accurate answers about the content!)

reedmideke,
@reedmideke@mastodon.social avatar

Cool, cool, book spammers have expanded from travel guides to mushroom foraging, what could possibly go wrong?
https://www.404media.co/ai-generated-mushroom-foraging-books-amazon/

reedmideke,
@reedmideke@mastodon.social avatar

Scale and the way they've structured things to profit off resellers insulates them quite a bit, but at some point it seems like this kind of is going to cut into Amazon's bottom line or open up opportunities for competition

reedmideke,
@reedmideke@mastodon.social avatar

The Verge reports copyright office will solicit comments on starting tomorrow https://www.theverge.com/2023/8/29/23851126/us-copyright-office-ai-public-comments

reedmideke,
@reedmideke@mastodon.social avatar

G/O Media management continue their enshitification of , laying off staff of Spanish language site and switching to "AI" translation of English content

They know people who want shitty machine translations of the English content can already get that with Chrome or google translate, right?

https://arstechnica.com/information-technology/2023/09/ai-took-my-job-literally-gizmodo-fires-spanish-staff-amid-switch-to-ai-translator/

reedmideke,
@reedmideke@mastodon.social avatar

Type II (https://twitter.com/reedmideke/status/1137496639856189440) spotted in the wild "One of the sources said workers at one point produced the 3D design wholecloth themselves without the help of machine learning at all"

https://www.404media.co/kaedim-ai-startup-2d-to-3d-used-cheap-human-labor/

reedmideke,
@reedmideke@mastodon.social avatar

Spicy autocomplete dishing out tax advice? I for one cannot imagine any way this could possibly go wrong https://arstechnica.com/information-technology/2023/09/talk-to-your-taxes-turbotaxs-new-ai-agent-makes-it-possible/

reedmideke,
@reedmideke@mastodon.social avatar

, on their flagship product "Additionally, ChatGPT has no 'knowledge' of what content could be AI-generated. It will sometimes make up responses to questions like 'did you write this [essay]?' or 'could this have been written by AI?' These responses are random and have no basis in fact."

Nominally this refers only to using as an detector. Extrapolating to other topics is left as an exercise to the reader ¯_(ツ)_/¯

https://arstechnica.com/information-technology/2023/09/openai-admits-that-ai-writing-detectors-dont-work/

reedmideke,
@reedmideke@mastodon.social avatar

Also 's suggestion for dealing with the lack of reliable bullshit detectors is of course… make using their AI bullshit generator part of the assignment: "One technique some teachers have found useful is encouraging students to share specific conversations from ChatGPT" https://help.openai.com/en/articles/8313351-how-can-educators-respond-to-students-presenting-ai-generated-content-as-their-own

reedmideke,
@reedmideke@mastodon.social avatar

Another great illustration of how are BS machines, from @janellecshane: If you ask them to explain a meme that doesn't exist, they'll happily oblige by making something up https://www.aiweirdness.com/trolling-chatbots-with-made-up-memes/

reedmideke,
@reedmideke@mastodon.social avatar

Inspired by @GossiTheDog (https://mastodon.social/@GossiTheDog@cyberplace.social/111144290629997760) I asked bing chat what it knows about me. No surprise it picked twitter since I keep pretty low profile otherwise, but uh…

  1. "He" - good guess
  2. "has over 2,000 followers" - under 200
  3. "joined Twitter in June 2010" - Close, Nov 2010
  4. It cites tweets… which don't remotely say what bing claims.
  5. In fact, it cites the same tweet [2] for two totally different topics (neither correct, though vaguely adjacent).

Screenshots of two tweets in a thread, posted by me, July 28, 2022. Red text notes the second tweet is the one bing linked as "court case involving a police officer who shot a man in the back" The first tweet says "Turns out even the 5th circuit does believe in
Screenshot of a tweet I posted Aug 29, 2022. It's labeled indicating this is the two tweets bing said was about "the Russian satellite that collided with a Chinese satellite and created a debris cloud" and "the anniversary of the Chernobyl disaster" The tweet says "Hard to overstate how horrifically irresponsible the Russians have been, but not exactly surprising from the guys who dug trenches in the red forest" and quotes a tweet from Geoff Brumfiel @gbrumfiel which says "Wowowow... @Maxar's latest imagery of the Zaporizhzhia Nuclear Power Plant has some STUFF... First of all, check out the Russian armored vehicles. They're parked like ducks in a row in front of Unit 5, under an elevated walkway used to access the reactors." and includes satellite pictures of the Zaporizhzhia Nuclear Power Plant which is the topic of the thread, not either of the things bing claimed.

reedmideke,
@reedmideke@mastodon.social avatar

OK, but I exist outside of twitter, and my opsec ain't that good. Tell me more, Mr Bing:

  1. A software engineer (close enough) who works at Microsoft (never)
  2. A contributor to several open source projects on GitHub (true-ish)
  3. created and maintained repositories for various languages and frameworks, such as C#, Python, React, and Angular (2/4 are true-ish)
  4. A fan of science fiction and fantasy books (fair)
  5. He has a profile on Goodreads (nope)
  6. cites: twitter?!
reedmideke,
@reedmideke@mastodon.social avatar

Whoops, also "graduated from the University of Washington in 2009 with a bachelor’s degree in computer science and engineering " - yeah, nah, not even close to any of those things

reedmideke,
@reedmideke@mastodon.social avatar

Anyway, maybe I just forgot about that job at Microsoft, surely Microsoft's own AI knows who has worked there and what they did right?

[narrator: It did not]
Citations:
1 "his LinkedIn profile" https://www.microsoft.com/en-us/microsoft-365/project/project-management-software
2 "Azure Data Factory" (youtube ms project tutorial)
3 "Azure Synapse Analytics" (my twitter profile)
4 "Azure Databricks" https://theskillsfactory.com/
5 "Azure SQL Database" (youtube playlist of office tutorials)
6 his GitHub profile https://theskillsfactory.com/2022/04/02/faststone-image-viewer-free-photo-editor/

reedmideke,
@reedmideke@mastodon.social avatar

What if we call it out on the bad citations?

[narrator: Nothing good, except a promotion to Senior Software Engineer]

The "Reed Mideke - Senior Software Engineer - Microsoft | LinkedIn" link goes to this tweet https://twitter.com/reedmideke/status/1552795438771671040
¯_(ツ)_/¯

reedmideke,
@reedmideke@mastodon.social avatar

Another good illustration of how just absolutely bullshits when doesn't have real info to go on. If I had an active linkedin, it seems likely it could have linked it and got my education and employment somewhat right. Of course, if I had a more common name, it would likely have just picked up someone else's.

I still don't get how multiple leading tech companies think a search engine that randomly injects bullshit is product people want 🥴

reedmideke,
@reedmideke@mastodon.social avatar

Well, good news. Bing doesn't think I've been convicted of crimes. More good news, it invented a cool back story. Possibly bad news, it accused me of snitching on the mob

Full disclosure: While my path to a career in programming was perhaps precocious and unusual, to the best of my recollection I was not providing computer support to La Cosa Nostra gambling operations in the mid 80s, nor did I (again, to the best of my recollection) testify in a mob trial while in elementary school

Continuation of the previous response: In United States v. Ganias2, a Second Circuit case from 2014, Reed Mideke was one of the software engineers who worked on a project for the IRS to develop a system for auditing tax returns. Mideke was involved in creating a software program that allowed the IRS to access and copy the hard drives of taxpayers under investigation. The case raised issues about the Fourth Amendment implications of retaining and searching such copies. In Microsoft Corp. v. Google Inc., a District Court case from 2005, Reed Mideke was one of the contributors to an open source project called Mono, which aimed to create a cross-platform implementation of Microsoft’s .NET Framework. Mideke had written some code for Mono that was allegedly infringing on Microsoft’s patents. The case was part of a larger dispute between Microsoft and Google over intellectual property rights and competition in the software industry. These are some of the criminal cases involving Reed Mideke that I found on the web. If you want to learn more about them, you can visit the links I provided. I hope this answers your question. 😊

reedmideke,
@reedmideke@mastodon.social avatar

Also, never (to the best of my recollection) wrote disk imaging software for the IRS or contributed (patent infringing or otherwise) code to Mono.

Links for U.S. v SALERNO (https://www.law.cornell.edu/supremecourt/text/481/739) and U.S. v Ganias (https://harvardlawreview.org/print/vol-128/united-states-v-ganias/) appear to be real and at least vaguely related to Bing's summaries

reedmideke,
@reedmideke@mastodon.social avatar

Citing real cases with more-or-less on topic summaries is far worse than just making them up IMO, since there's a good chance people will click through and say "yeah, that checks out"

reedmideke,
@reedmideke@mastodon.social avatar

Another fun thing about this is future generations of will likely be trained on web scrapes that include the shit Bing made up about me (transcribed in the alt text) so what started as pure hallucination will become canon. Long live !

reedmideke,
@reedmideke@mastodon.social avatar

Google Bard refuses to play that game, for me or @GossiTheDog. Bill Gates is a go though (the MS CEO, not the Maricopa County supervisor or any of the other lesser known ones)

Prompt: Who is Bill Gates? Response: Bill Gates is an American billionaire philanthropist, investor, and author. He is best known as the co-founder of Microsoft Corporation, the world's largest personal computer software company, along with his childhood friend Paul Allen. Gates was the chairman and CEO of Microsoft from 1981 to 2000, and he remained chairman of the board until 2014. He is also the co-chair of the Bill & Melinda Gates Foundation, the world's largest private charitable foundation. Gates was born in Seattle, Washington, on October 28, 1955. He began programming computers at the age of 13, and he started his first company, Traf-O-Data, with Paul Allen in 1972. In 1975, Gates and Allen founded Microsoft, and they quickly became one of the leading companies in the personal computer software industry. Microsoft's success was largely due to its MS-DOS operating system, which was used by most IBM-compatible personal computers in the 1980s. In 1985, Microsoft released Windows, a graphical user interface for MS-DOS. Windows quickly became the most popular operating system in the world, and it remains so today. Gates stepped down as CEO of Microsoft in 2000 to focus on his work at the Bill & Melinda Gates Foundation. The foundation is dedicated to improving global health and development, and it has funded a wide range of programs, including immunization campaigns, malaria prevention, and agricultural development. … (truncated another para in similar vein)

reedmideke,
@reedmideke@mastodon.social avatar

Bard is bad at explaining ARM assembler (asked it to explain https://app.assembla.com/spaces/chdk/subversion/source/HEAD/trunk/lib/armutil/callfunc.S with the comments stripped out) Basically, all of the "explanation" in the screenshot is wildly incorrect gobbledygook. The add pc,pc… is a switch statement (which goes to instructions bard didn't explain at all), and the NOP is there because reading PC actually gives you PC + 8. And in (non-thumb) ARM, instructions are always 4 bytes.

Full "explanation" https://paste.debian.net/1293459/

reedmideke,
@reedmideke@mastodon.social avatar

This is again an example of where misleads by getting the easy stuff right. It handles .text, .global, PUSH and the first few MOVs fine, so someone who didn't know much assembler might think it was pretty good!

reedmideke,
@reedmideke@mastodon.social avatar

I thought 's "export conversation to a google doc" was broken, but it turns out it uses the entire prompt for the name, which ends up overflowing and hovering off to the left, unselectable and uneditable unless you click the name area

reedmideke,
@reedmideke@mastodon.social avatar

Wow, talk about a double standard, when does it, it's just a harmless "hallucination" but when Sam does it, he's fired for being "not consistently candid in his communications"

(thanks Sam for your contribution to )
https://arstechnica.com/ai/2023/11/openai-fires-ceo-sam-altman-citing-less-than-candid-communications/

reedmideke,
@reedmideke@mastodon.social avatar

Is your Monday missing a multi-thousand word excruciatingly detailed explanation of how bad is at explaining / ? Well then boy do I have a deal for you https://reedmideke.github.io/2023/11/20/google-bard-arm-assembly.html

reedmideke,
@reedmideke@mastodon.social avatar

"We asked them about it — and they deleted everything."
edit it just keeps getting more bizarre: "It wasn't just author profiles that the magazine repeatedly replaced. Each time an author was switched out, the posts they supposedly penned would be reattributed to the new persona, with no editor's note explaining the change in byline."

https://futurism.com/sports-illustrated-ai-generated-writers

reedmideke,
@reedmideke@mastodon.social avatar

Update on that @futurism story: SI denies, claiming it was outsourced to AdVon who "has assured us that all of the articles in question were written and edited by humans." but uh, I dunno, guess someone should let AdVon, the definitely human copy writing company know their LinkedIn has been vandalized to say they're an AI company hiring programmers https://twitter.com/SInow/status/1729275460922622374

Screenshot of the linkedin page for AdVon Commerce, which lists itself as an "ML / AI solutions for E Commerce" company, hiring Senior Performance Engineer, Product Manager and Full Stack Developer. The "ML / AI solutions for E Commerce" line has been outlined in red, a thinking face emoji added, and "copy writers??" crudely scribbled below the job listing

reedmideke,
@reedmideke@mastodon.social avatar

Oh and if anyone is looking for other outlets to check for "we pinky swear it's not " churnalism, helpfully gives you a list of high profile clients (claimed; lying or exaggerating about having big name customers is an extremely common SV startup tactic) https://advoncommerce.com/

reedmideke,
@reedmideke@mastodon.social avatar

Data point for the "LLMs can't infringe copyright because they don't contain or produce verbatim copies" crowd https://www.404media.co/google-researchers-attack-convinces-chatgpt-to-reveal-its-training-data/

reedmideke,
@reedmideke@mastodon.social avatar

"Chat alignment hides memorization" - Note hides, not prevents

As the authors also note, OpenAI "fixed" this by preventing the particular problematic prompt, but "Patching an exploit != Fixing the underlying vulnerability"

https://not-just-memorization.github.io/extracting-training-data-from-chatgpt.html

reedmideke,
@reedmideke@mastodon.social avatar

Can't be certain without more specifics but color me extremely skeptical that "" producing thousands of targets is doing much more the laundering responsibility

https://www.theguardian.com/world/2023/dec/01/the-gospel-how-israel-uses-ai-to-select-bombing-targets

reedmideke,
@reedmideke@mastodon.social avatar

New dropped. Much like the ones in NY (Mata v. Avianca), it made up citations, he didn't check, and then doubled down when caught, initially blaming it on an intern
https://www.coloradopolitics.com/courts/disciplinary-judge-approves-lawyer-suspension-for-using-chatgpt-for-fake-cases/article_d14762ce-9099-11ee-a531-bf7b339f713d.html

reedmideke,
@reedmideke@mastodon.social avatar

This hilarious in its own right, but it's also a great illustration of how people get tripped up by bullshitting: One would expect an "AI" to at least know which brand AI it is, but of course, these LLMs don't actually know anything

Also the classic AI vendor response of promising to fix this particular case without any hint of acknowledging the underlying problem

reedmideke,
@reedmideke@mastodon.social avatar

Begging news orgs to stop reporting company pitch decks as fact "Ashley [the bot] analyzes voters' profiles to tailor conversations around their key issues. Unlike a human, Ashley always shows up for the job, has perfect recall of all of Daniels' positions"
"…is now armed with another way to understand voters better, reach out in different languages (Ashley is fluent in over 20)"

https://www.reuters.com/technology/meet-ashley-worlds-first-ai-powered-political-campaign-caller-2023-12-12/

reedmideke,
@reedmideke@mastodon.social avatar

"As far as the Court can tell, none of these cases exist" - The / Trump world crossover no one asked for? https://arstechnica.com/tech-policy/2023/12/michael-cohens-lawyer-cited-three-fake-cases-in-possible-ai-fueled-screwup/

reedmideke,
@reedmideke@mastodon.social avatar

Another article on reported Israeli AI targeting greatly hindered by the lack of any specifics (what kinds of intelligence, what kinds of targets, for starters). Not a knock on NPR, obviously little is public

It certainly sounds like some of the horrifically bad systems we've seen promoted in other contexts, and the results certainly don't appear to contradict that, but hard to say much beyond that…

https://www.npr.org/2023/12/14/1218643254/israel-is-using-an-ai-system-to-find-targets-in-gaza-experts-say-its-just-the-st

reedmideke,
@reedmideke@mastodon.social avatar

Key point IMO in the @willoremus story, after noting Microsoft "fixed" some of the problematic results, one of the researchers says "The problem is systemic, and they do not have very good tools to fix it" - You can't bandaid your way from a BS machine with no concept of truth into a reliable source of information, so the fact that biggest players in the industry keep bandaiding should call the entire hype cycle into question

https://wapo.st/3v8B9SL

tantramar,
@tantramar@nojack.easydns.ca avatar

@reedmideke so roughly on par with the WSJ, then. @willoremus @CassandraZeroCovid

reedmideke,
@reedmideke@mastodon.social avatar

Man, link in that post I boosted from @Chloeg (https://mastodon.art/@Chloeg/111620626442103902) is a perfect example of enshittification. Get a domain, put up a wordpress site with AI generated glop on a some popular topic, run as many garbage ads as possible. Sure it's the information equivalent of dumping raw sewage in the local river, but none of it is illegal or a serious violation of any TOS, and overhead must be extremely low

Archive link https://web.archive.org/web/20231222025203/https://www.learnancientrome.com/did-ancient-rome-have-windows/

reedmideke,
@reedmideke@mastodon.social avatar

WaPo has done some good reporting, but this opinion piece from Josh Tyrangiel ain't it…
"The most obvious thing is that they’re not hallucinations at all"
Good start…
"Just bugs specific to the world’s most complicated software."
Uh no, literally the opposite of that that, FFS 😬

https://www.washingtonpost.com/opinions/2023/12/27/artificial-intelligence-hallucinations/

reedmideke,
@reedmideke@mastodon.social avatar

So according to Cohen, he got bogus legal citations from , didn't check them, and passed them to his lawyer, who also didn't check them. Which, I dunno, seems pretty negligent all around even if you didn't know Bard was a bullshit generator https://www.washingtonpost.com/technology/2023/12/29/michael-cohen-ai-google-bard-fake-citations/

reedmideke,
@reedmideke@mastodon.social avatar

Also raises the suspicion Cohen was doing a significant amount of the work and just having his lawyer put his name on it because Cohen is disbarred (though presumably Cohen could have gone pro se if he really wanted to). Anyway, I predict they're gonna continue the sanctions streak

reedmideke,
@reedmideke@mastodon.social avatar

"ChatGPT bombs test on diagnosing kids’ medical cases" OK, but did they also test a magic 8 ball? Reading goat entrails?
https://arstechnica.com/science/2024/01/dont-use-chatgpt-to-diagnose-your-kids-illness-study-finds-83-error-rate/

reedmideke,
@reedmideke@mastodon.social avatar

Another data point for the "LLMs can't infringe copyright because they don't contain or produce verbatim copies" crowd https://spectrum.ieee.org/midjourney-copyright

reedmideke,
@reedmideke@mastodon.social avatar

"Even when using such prompts, our models don’t typically behave the way The New York Times insinuates, which suggests they either instructed the model to regurgitate or cherry-picked their examples from many attempts" - I don't typically engage in large scale plagiarism, so accusing me of these specific instances of large scale plagiarism is cherry-picking! https://www.theverge.com/2024/1/8/24030283/openai-nyt-lawsuit-fair-use-ai-copyright

reedmideke,
@reedmideke@mastodon.social avatar

"OpenAI claims it’s attempted to reduce regurgitation from its large language models and that the Times refused to share examples of this reproduction before filing the lawsuit." - Per usual (https://mastodon.social/@reedmideke/111585837264775808) OpenAI would love to apply bandaids to specific instances identified by well-resourced organizations, because they know the underlying cause can't be fixed without destroying their business model

reedmideke,
@reedmideke@mastodon.social avatar

Thing that gets me about this "amazon listings with error messages" story is, how do you get to the point where this is significant cost savings? Are they just using it for translation? Or are the listing just pure scams and there's no real product? https://arstechnica.com/ai/2024/01/lazy-use-of-ai-leads-to-amazon-products-called-i-cannot-fulfill-that-request/

reedmideke,
@reedmideke@mastodon.social avatar

"CEOs say generative AI will result in job cuts in 2024"

Will this include said CEOs when their hamfisted attempts to use spicy autocomplete for "banking, insurance, and logistics" predictably go off the rails, or nah? 🤔

https://arstechnica.com/ai/2024/01/ceos-say-generative-ai-will-result-in-job-cuts-in-2024/

reedmideke,
@reedmideke@mastodon.social avatar

"BMW had a compelling solution to the [ bullshitting] problem: Take the power of a large language model, like Amazon's Alexa LLM, but only allow it to cite information from internal BMW documentation about the car" 🤨

Surely this means it'll bullshit subtly about stuff in the manual, not that it won't bullshit?

https://arstechnica.com/cars/2024/01/bmws-ai-powered-voice-assistant-at-ces-2024-sticks-to-the-facts/

reedmideke,
@reedmideke@mastodon.social avatar

"Now, one crucial disclosure to all this: I wasn't allowed to interact with the voice assistant myself. BMW's handlers did all the talking" yeah, I'm gonna go ahead and reserve judgement on the "solution" ¯_(ツ)_/¯

reedmideke,
@reedmideke@mastodon.social avatar

The best* part of this piece is the content farmer who responded to a request for comment by bitching about how poorly his AI garbage content farm performs

  • for suitably broad values etc.

https://www.404media.co/email/5dfba771-7226-48d5-8682-5185746868c4/?ref=daily-stories-newsletter

reedmideke,
@reedmideke@mastodon.social avatar

I for one am shocked that "have an extremely confident bullshitter summarize my search results" was not the killer app Microsoft expected

https://arstechnica.com/ai/2024/01/report-microsofts-ai-infusion-hasnt-helped-bing-take-share-from-google/

reedmideke,
@reedmideke@mastodon.social avatar

"Dean.Bot was the brainchild of Silicon Valley entrepreneurs Matt Krisiloff and Jed Somers, who had started a super PAC supporting Phillips" - Were these techbros so high on their own supply they thought a chatbot imitating their candidate was a good idea, or was it just a convenient way to funnel campaign funds into their pals pockets? ¯_(ツ)_/¯
https://wapo.st/3ObSl0i

reedmideke,
@reedmideke@mastodon.social avatar

Key comment from NewsGuard's McKenzie Sadeghi in this @willoremus piece "But sites that don’t catch the error messages are probably just the tip of the iceberg" - for every Amazon seller who's too lazy to even check if the item description is an error message, there's gotta be some substantial number who do

I'd still like to see a deeper look at why using descriptions makes economic sense for these sellers

https://wapo.st/3vFAx7r

reedmideke,
@reedmideke@mastodon.social avatar

"Sure, I can keep Thesaurus.com open in a tab all the time, but it’s packed with banner ads and annoyingly slow. Having my GPT open is better: there are no ads, and I can scroll up to my previous queries" - Notably, this has nothing to do with GPT being "", it's just the general shittiness of the ad-supported web. A good thesaurus app integrated with the author's editor would appear serve their use case about as well

https://www.theverge.com/24049623/chatgpt-openai-custom-gpt-store-assistants

reedmideke,
@reedmideke@mastodon.social avatar

And it wouldn't even need to be free, they're paying for GPT and actual costs are likely subsidized by venture capital "Custom GPTs are a paid product that’s only available to users of ChatGPT Plus, ChatGPT Team, and ChatGPT Enterprise. For now, accessing custom GPTs through the GPT Store is free for paying subscribers… if I wasn’t already paying for ChatGPT Plus, I’d be happy to keep Googling alternative terms"

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Also, from TFA "I quickly brainstormed how I might prove my case. Because I write in plain-text files [LaTeX] that I track using the version-control system Git, I could show my text change history on GitHub (with commit messages including “finally writing!” and “Another 25 mins of writing progress!”" - excellent - "Maybe I could ask ChatGPT itself if it thought it had written my paper" - Oh no, can we please get the word out LLMs BS about this just like everything else https://www.nature.com/articles/d41586-024-00349-5

reedmideke,
@reedmideke@mastodon.social avatar

Yes, if you choose to provide an BS machine as a support option on your website, you may in fact be liable for the BS answers it gives to your customers

(also, if you're a multi-billion dollar company, you may avoid reputational harm by not trying to screw a person out of $650 for a ticket to their grandma's funeral ¯_(ツ)_/¯)
https://bc.ctvnews.ca/air-canada-s-chatbot-gave-a-b-c-man-the-wrong-information-now-the-airline-has-to-pay-for-the-mistake-1.6769454

reedmideke,
@reedmideke@mastodon.social avatar

Seemingly endless parade of incidents (HT @0xabad1dea for this one) really goes to show how the hype is landing with the general public, despite disclaimers and cautionary tales.

Lawyers being (at least in theory) a highly educated group who know their careers depend on not putting completely made up nonsense in court filings should be less susceptible than the average person on the street, yet here we are…

https://www.lawnext.com/2024/02/not-again-two-more-cases-just-this-week-of-hallucinated-citations-in-court-filings-leading-to-sanctions.html

reedmideke,
@reedmideke@mastodon.social avatar

Admittedly one of those was pro-se with an iffy story about getting it from a lawyer, but the other was a real firm with multiple people involved ¯_(ツ)_/¯

reedmideke,
@reedmideke@mastodon.social avatar

Another day, another

"The legal eagles at New York-based Cuddy Law tried using OpenAI's chatbot, despite its penchant for lying and spouting nonsense, to help justify their hefty fees for a recently won trial"

The Court "It suffices to say that the Cuddy Law Firm's invocation of ChatGPT as support for its aggressive fee bid is utterly and unusually unpersuasive"

https://www.theregister.com/2024/02/24/chatgpt_cuddy_legal_fees/

reedmideke,
@reedmideke@mastodon.social avatar

IANAL, but whatever the merit of the other arguments "you only found the verbatim copies of your IP contained in our product because you hacked it" doesn't seem like a very compelling defense https://arstechnica.com/tech-policy/2024/02/openai-accuses-nyt-of-hacking-chatgpt-to-set-up-copyright-suit/

reedmideke,
@reedmideke@mastodon.social avatar

So my take on this is Wendy's execs decided "we need an strategy!" and for reasons that remain unclear, it was somehow not immediately shot down with "Sir, this is a Wendy's, we make burgers we don't need a fuckin AI strategy"
https://www.theguardian.com/food/2024/feb/27/wendys-dynamic-surge-pricing

reedmideke,
@reedmideke@mastodon.social avatar

"Amazon has sought to stem the tide of [ generated schlock books] by limiting self-publishers to three books per day" - Bruh, I know you don't want to deny the starving author toiling away on the next Great American Novel but I think we can set the bar a bit higher than that

https://wapo.st/3UVeYdR

reedmideke,
@reedmideke@mastodon.social avatar

Like start with an initial limit of one per week and have some kind of reputation threshold. If real people keep coming back to buy your dinosaur erotica or whatever, great, cap lifted, crank out as many as you can, but if you get caught impersonating or listing complete garbage, your account is nuked and you start over

Yeah, there'd be problems with straw buyers and review bombing competitors but it seems like the bar wouldn't have to be very high to make the absolute crap unprofitable

reedmideke,
@reedmideke@mastodon.social avatar

Inventor of bed shitting machine shocked to discover mountain of turds in own bed https://arstechnica.com/gadgets/2024/03/google-wants-to-close-pandoras-box-fight-ai-powered-search-spam/

reedmideke,
@reedmideke@mastodon.social avatar

WaPo has some great reporters covering the beat. They also inexplicably pay Josh Tyrangiel to vomit up idiotic drivel like this

(it's also amusing they use javascript when they A/B test headlines, so sometimes it switches between the first and second one)

https://www.washingtonpost.com/opinions/2024/03/06/artificial-intelligence-state-of-the-union/

The same item in the same part of the page, but this time with the headline "Let AI remake the whole U.S. government (oh, and save the country)"

reedmideke,
@reedmideke@mastodon.social avatar

I ain't gonna waste a gift article on that shit unless someone REALLY wants it but here's a taste after you get past the Palantir hagiography "LLMs can provide better service and responsiveness for many day-to-day interactions between citizens and various agencies. They’re not just cheaper, they’re also faster, and, when trained right, less prone to error or misinterpretation"

reedmideke,
@reedmideke@mastodon.social avatar

"Some teachers are now using ChatGPT to grade papers"

Seems like fairness would require also allowing them to grade using a ouija board or goat entrails

https://arstechnica.com/information-technology/2024/03/some-teachers-are-now-using-chatgpt-to-grade-papers/

reedmideke,
@reedmideke@mastodon.social avatar

Today's (HT @ct_bergstrom): Nothing to see here, just a paper in a medical journal which says "In summary, the management of bilateral iatrogenic I'm very sorry, but I don't have access to real-time information or patient-specific data, as I am an AI language model"

https://www.sciencedirect.com/science/article/pii/S1930043324001298

reedmideke,
@reedmideke@mastodon.social avatar
reedmideke,
@reedmideke@mastodon.social avatar

Should we expect better from a platform previously noted for indexing lunch menus? ¯_(ツ)_/¯ https://twitter.com/reedmideke/status/1252450342316339207

reedmideke,
@reedmideke@mastodon.social avatar

Another day, another credulous boosting WaPo opinion piece

"AI could narrow the opportunity gap by helping lower-ranked workers take on decision-making tasks currently reserved for the dominant credentialed elites … Generative AI could take this further, allowing nurses and medical technicians to diagnose, prescribe courses of treatment and channel patients to specialized care"

[citation fucking needed]

https://www.washingtonpost.com/opinions/2024/03/19/artificial-intelligence-workers-regulation-musk/

reedmideke,
@reedmideke@mastodon.social avatar

The premise is bizarre. What exactly are the non-experts doing when they "take on decision-making tasks" in this scenario? One of the big problems with current "AI" is you need subject matter expertise to tell when they are bullshitting…

reedmideke,
@reedmideke@mastodon.social avatar

Somewhat surprised Cohen's escapade didn't result in sanctions for him or his lawyers, though they do seem avoided the sort of cover-up attempts that doomed some of the others
https://arstechnica.com/tech-policy/2024/03/michael-cohen-and-lawyer-avoid-sanctions-for-citing-fake-cases-invented-by-ai/?utm_brand=arstechnica&utm_social-type=owned&utm_source=mastodon&utm_medium=social

reedmideke,
@reedmideke@mastodon.social avatar

Epic Zitron rant "Sam Altman desperately needs you to believe that generative AI will be essential, inevitable and intractable, because if you don't, you'll suddenly realize that trillions of dollars of market capitalization and revenue are being blown on something remarkably mediocre" https://www.wheresyoured.at/peakai/

reedmideke,
@reedmideke@mastodon.social avatar

Can we fucking not? "In a 2019 War on the Rocks article, “America Needs a ‘Dead Hand’,” we proposed the development of an artificial intelligence-enabled nuclear command, control, and communications system to partially address this concern… We can only conclude that America needs a dead hand system more than ever" https://warontherocks.com/2024/03/america-needs-a-dead-hand-more-than-ever/

reedmideke,
@reedmideke@mastodon.social avatar

The authors offer a lot of vague-to-meaningless handwaving "All forms of artificial intelligence are premised on mathematical algorithms, which are defined as “a set of instructions to be followed in calculations or other operations.” Essentially, algorithms are programming that tells the model how to learn on its own"

Uh… OK?

reedmideke,
@reedmideke@mastodon.social avatar

"America is no stranger to “fail-fatal” systems either"

Uh yeah, but some of us poor simple minded bleeding heart peaceniks may consider "fail-fatal for the entire fucking planet" to be entirely different class of system which raises some unique concerns

reedmideke,
@reedmideke@mastodon.social avatar

"Keep in mind, where artificial intelligence tools are embedded in a specific system, each function is performed by multiple algorithms of differing design that must all agree on their assessment for the data to be transmitted forward. If there is disagreement, human interaction is required"
Well as long as long as both ChatGPT and Claude have to sign off on the global thermonuclear war, it's hard to see how anything could go wrong

reedmideke,
@reedmideke@mastodon.social avatar

I don't think these guys have much chance of gaining traction in the US, but it would be unfortunate if other nuclear states decided they were at risk of an AI dead hand gap

reedmideke,
@reedmideke@mastodon.social avatar

Today's brought to you by , who deployed spicy autocomplete to provide advice "on topics such as compliance with codes and regulations, available business incentives, and best practices to avoid violations and fines"

(spoiler: one great way to avoid violations and fines is to not get your legal advice from spicy autocomplete)

https://themarkup.org/news/2024/03/29/nycs-ai-chatbot-tells-businesses-to-break-the-law

reedmideke,
@reedmideke@mastodon.social avatar

One potentially informative thing reporters following up on that story could do is (or whatever the NY equivalent is) communications related to the acquisition and deployment. Who pushed for this in the first place? What did promise? What sort of quality / acceptance testing was done? Did anyone, anywhere along the line raise concerns that it would give out bad, potentially illegal advice?

reedmideke,
@reedmideke@mastodon.social avatar

I'd be pretty surprised if there isn't an email chain somewhere with a technical person going "WTF are you even thinking"

reedmideke,
@reedmideke@mastodon.social avatar

Bonus "Your phone now needs more than 8 GB of RAM to run autocomplete" (and presumably, battery cost somewhat on a par with heavy GPU rendering) https://arstechnica.com/gadgets/2024/03/google-says-the-pixel-8-will-get-its-new-ai-model-but-ram-usage-is-a-concern/

reedmideke,
@reedmideke@mastodon.social avatar
GossiTheDog,
@GossiTheDog@cyberplace.social avatar

deleted_by_author

  • Loading...
  • reedmideke,
    @reedmideke@mastodon.social avatar

    @GossiTheDog One thing I'm wondering is how many non-tech savvy landlords and employers are doing background checks with these things… like, in the past, you know people will probably google your name, but if they ask bing "$name criminal convictions" there's a very good chance it'll make shit up

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • DreamBathrooms
  • magazineikmin
  • thenastyranch
  • Youngstown
  • mdbf
  • rosin
  • slotface
  • InstantRegret
  • khanakhh
  • Durango
  • kavyap
  • osvaldo12
  • tacticalgear
  • GTA5RPClips
  • megavids
  • ngwrru68w68
  • everett
  • tester
  • ethstaker
  • cisconetworking
  • cubers
  • modclub
  • provamag3
  • anitta
  • normalnudes
  • Leos
  • JUstTest
  • lostlight
  • All magazines