cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

BARD --> Gemini bumpiness shows the precarious position that LLM foundation models based on huge badly-managed datasets put companies in. Google is rushing. And the results are wrong and bad.

https://www.wsj.com/tech/ai/google-mired-in-controversy-over-ai-chatbot-push-46023dd3

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Just delivered the first BIML LLM Risks talk at NDSS in San Diego. Much fun was had!

Getting set up for the talk...

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

BIML believes that reproducibility economics is a serious risk to scientific study of .

"Mensch said his new model cost less than €20 million, to train. By contrast OpenAI Chief Executive Sam Altman said last year after the release of GPT-4 that training his company’s biggest models cost “much more than” $50 million to $100 million."

Which academic CS organizations have a "mere $22M" to build an LLM to experiment with. How can you try alternatives?

https://www.wsj.com/tech/ai/the-9-month-old-ai-startup-challenging-silicon-valleys-giants-ee2e4c48

cigitalgem, to security
@cigitalgem@sigmoid.social avatar

I will try to beat @0xmchow to the punch since it's my 58th birthday!

Secure your ML algorithms too while you're at it.

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Just finished hacking up slides for the LLM security work BIML recently released. I will be presenting this invited talk for three NDSS conference workshops (simultaneously) in San Diego Monday afternoon.

All NDSS ’24 workshops: https://www.ndss-symposium.org/ndss2024/co-located-events/

  1. SDIoTSec: https://www.ndss-symposium.org/ndss2024/co-located-events/sdiotsec/
  2. USEC: https://www.ndss-symposium.org/ndss2024/co-located-events/usec/
  3. AISCC: https://www.ndss-symposium.org/ndss2024/co-located-events/aiscc/.
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

As a pizza delivery person you too can prompt persnickety parrots with pen test panache using this new tool from Microsoft. A whole new cyber cyber career!

https://www.microsoft.com/en-us/security/blog/2024/02/22/announcing-microsofts-open-automation-framework-to-red-team-generative-ai-systems/

I know, let's pretend that LLM security can be bolted on later after we have created a foundation model based on data scraped from the Internet that is FULL of poison, garbage, nonsense, and noise. <Announcer: It can't>

cigitalgem, to random
@cigitalgem@sigmoid.social avatar

Building Security In can be done with LLM applications

https://blog.redsift.com/news/announcing-red-sift-radar-beta/

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

NEW Security Ledger podcast features BIML's LLM risk analysis, recursive pollution, and data feudalism. Always a great time chatting with Paul Roberts! @securityledger

https://securityledger.com/2024/02/episode-256-recursive-pollution-data-feudalism-gary-mcgraw-on-llm-insecurity/

cigitalgem,
@cigitalgem@sigmoid.social avatar

The biggest risk posed by large language model AI like Chat GPT? “It’s this: large language models are often wrong,” McGraw told me. “And they’re very convincingly wrong and very authoritatively wrong.”

cigitalgem, to random
@cigitalgem@sigmoid.social avatar

"AI" (that is black box auto-associative ML generators) will ensnare your attention and not let go. Who needs links on the web to original content if an ML copycat can condense it all for you---introducing errors and nonsense while doing so??

https://www.fastcompany.com/91033052/does-anyone-even-want-an-ai-search-engine

cigitalgem, to random
@cigitalgem@sigmoid.social avatar
cigitalgem, to random
@cigitalgem@sigmoid.social avatar

Wide ranging BIML today discussing LLMs and the law as well as LLM Evaluation. The works we read were not remarkable.

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

When ML goes wrong, who pays the price?

Includes 3 of 81 relevant BIML LLM risks from our 1.24.24 report.

https://berryvilleiml.com/2024/02/15/when-ml-goes-wrong-who-pays-the-price/

cigitalgem, (edited ) to random
@cigitalgem@sigmoid.social avatar
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar
cigitalgem, to ML
@cigitalgem@sigmoid.social avatar
cigitalgem, to random
@cigitalgem@sigmoid.social avatar

Thank goodness for Carlini's work. A breath of fresh air after the vendor nonsense.

cigitalgem, to random
@cigitalgem@sigmoid.social avatar

The "sleeper agents" paper from anthropic is such a complete bullshit I don't even know where to start. Good grief...such terrible "science."

cigitalgem,
@cigitalgem@sigmoid.social avatar

BIML reviewed this terrible work yesterday. It is so badly done that we feel the need to respond to it. This is the second anthropic paper on AI alignment that presented thin gruel, poor reasoning, and a misunderstanding of the basics of science. That kind of work does nothing to advance the field of .

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Have a listen to BIML discuss Machine Learning Security on the Google Cloud Security podcast

https://berryvilleiml.com/2024/01/25/google-cloud-security-podcast-features-biml/

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Here's another example of recursive pollution. Garbage news which then gets ingested.

https://www.404media.co/google-news-is-boosting-garbage-ai-generated-articles/

Wonder why this is terrible? Check out BIML's LLM Risk Analysis
https://berryvilleiml.com/results/

cigitalgem, to ML
@cigitalgem@sigmoid.social avatar

Lazy use of ML to generate product descriptions automatically. Guess what, this is a prime example of recursive pollution...because these descriptions will be eaten by search engines, etc. Here we go!


https://arstechnica.com/ai/2024/01/lazy-use-of-ai-leads-to-amazon-products-called-i-cannot-fulfill-that-request/

  • All
  • Subscribed
  • Moderated
  • Favorites
  • provamag3
  • magazineikmin
  • Youngstown
  • osvaldo12
  • khanakhh
  • slotface
  • tacticalgear
  • InstantRegret
  • ngwrru68w68
  • kavyap
  • DreamBathrooms
  • thenastyranch
  • everett
  • rosin
  • JUstTest
  • Durango
  • GTA5RPClips
  • ethstaker
  • modclub
  • mdbf
  • cisconetworking
  • Leos
  • normalnudes
  • cubers
  • megavids
  • tester
  • anitta
  • lostlight
  • All magazines