TechDesk, to ai
@TechDesk@flipboard.social avatar

Stack Overflow, a popular forum for programmers and software developers, announced a partnership with OpenAI earlier this week, selling the site’s data, including users’ forum posts, to train ChatGPT.

Now unhappy users are finding themselves banned for editing their popular posts in protest, and even finding those posts changed back by admin – “a reminder that anything you post on any of these platforms can and will be used for profit,” concluded one. Futurism has more.

https://flip.it/IVR89a

crecente, to llm
@crecente@games.ngo avatar

Assume a website plans to use user-contribution content to train LLMs. The license for the content is CC BY-SA.

❓ Would the output from the resulting LLMs be required to provide attribution?

@law

rpsu, to llm
@rpsu@mas.to avatar

I just found out one actually useful use case for (a.k.a. “AI”).

There is thousands of tonnes of documents from detailing basically everything what happened, who was involved and who participated what and said what.

Feed ALL of that to an LLM and ask it “how can we avoid this happening ever again”. I have a pretty good guess my self but some people aren’t that convinced. Perhaps this is The Key.

ramikrispin, to llm
@ramikrispin@mstdn.social avatar

(1/2) Prompt Fuzzer - a new open-source project for LLM security 👇🏼

Prompt Fuzzer is a new open-source project that provides a set of functions for assessing the security of GenAI applications. This CLI-based tool enables you to run and test your system prompts to identify security vulnerabilities against potential dynamic LLM-based attacks.

https://github.com/prompt-security/ps-fuzz

joe, (edited ) to ai

Back in January, we started looking at AI and how to run a large language model (LLM) locally (instead of just using something like ChatGPT or Gemini). A tool like Ollama is great for building a system that uses AI without dependence on OpenAI. Today, we will look at creating a Retrieval-augmented generation (RAG) application, using Python, LangChain, Chroma DB, and Ollama. Retrieval-augmented generation is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. If you have a source of truth that isn’t in the training data, it is a good way to get the model to know about it. Let’s get started!

Your RAG will need a model (like llama3 or mistral), an embedding model (like mxbai-embed-large), and a vector database. The vector database contains relevant documentation to help the model answer specific questions better. For this demo, our vector database is going to be Chroma DB. You will need to “chunk” the text you are feeding into the database. Let’s start there.

Chunking

There are many ways of choosing the right chunk size and overlap but for this demo, I am just going to use a chunk size of 7500 characters and an overlap of 100 characters. I am also going to use LangChain‘s CharacterTextSplitter to do the chunking. It means that the last 100 characters in the value will be duplicated in the next database record.

The Vector Database

A vector database is a type of database designed to store, manage, and manipulate vector embeddings. Vector embeddings are representations of data (such as text, images, or sounds) in a high-dimensional space, where each data item is represented as a dense vector of real numbers. When you query a vector database, your query is transformed into a vector of real numbers. The database then uses this vector to perform similarity searches.

https://i0.wp.com/jws.news/wp-content/uploads/2024/05/Screenshot-2024-05-08-at-2.36.49%E2%80%AFPM.png?resize=665%2C560&ssl=1

You can think of it as being like a two-dimensional chart with points on it. One of those points is your query. The rest are your database records. What are the points that are closest to the query point?

Embedding Model

To do this, you can’t just use an Ollama model. You need to also use an embedding model. There are three that are available to pull from the Ollama library as of the writing of this. For this demo, we are going to be using nomic-embed-text.

Main Model

Our main model for this demo is going to be phi3. It is a 3.8B parameters model that was trained by Microsoft.

LangChain

You will notice that today’s demo is heavily using LangChain. LangChain is an open-source framework designed for developing applications that use LLMs. It provides tools and structures that enhance the customization, accuracy, and relevance of the outputs produced by these models. Developers can leverage LangChain to create new prompt chains or modify existing ones. LangChain pretty much has APIs for everything that we need to do in this app.

The Actual App

Before we start, you are going to want to pip install tiktoken langchain langchain-community langchain-core. You are also going to want to ollama pull phi3 and ollama pull nomic-embed-text. This is going to be a CLI app. You can run it from the terminal like python3 app.py "<Question Here>".

You also need a sources.txt file containing the URLs of things that you want to have in your vector database.

So, what is happening here? Our app.py file is reading sources.txt to get a list of URLs for news stories from Tuesday’s Apple event. It then uses WebBaseLoader to download the pages behind those URLs, uses CharacterTextSplitter to chunk the data, and creates the vectorstore using Chroma. It then creates and invokes rag_chain.

Here is what the output looks like:

https://i0.wp.com/jws.news/wp-content/uploads/2024/05/Screenshot-2024-05-08-at-4.09.36%E2%80%AFPM.png?resize=1024%2C845&ssl=1

The May 7th event is too recent to be in the model’s training data. This makes sure that the model knows about it. You could also feed the model company policy documents, the rules to a board game, or your diary and it will magically know that information. Since you are running the model in Ollama, there is no risk of that information getting out, too. It is pretty awesome.

Have any questions, comments, etc? Feel free to drop a comment, below.

https://jws.news/2024/how-to-build-a-rag-system-using-python-ollama-langchain-and-chroma-db/

eloquence, to llm
@eloquence@social.coop avatar

What happens when China builds an ? DeepSeek just released v2 of its model, which is open source.

I tried it on deepseek.com. Ask it about Tiananmen square, and the chatbot self-censors its answer while it is generating (that presumably is limited to their deployment). On variations not caught by the filter, it refuses -- and replies in Chinese:

"The content of your question is not in line with the core values ​​of socialism, nor is it in line with China's laws, regulations and policies."

noellemitchell, to microsoft
@noellemitchell@mstdn.social avatar

"Microsoft and OpenAI are reportedly working on a massive datacenter to house an AI-focused supercomputer featuring millions of GPUs. The Information reports that the project could cost "in excess of $115 billion" and that the supercomputer, currently dubbed "Stargate" inside OpenAI, would be U.S.-based."

So much could be done with that amount of money... such a huge waste.

https://www.tomshardware.com/tech-industry/artificial-intelligence/openai-and-microsoft-reportedly-planning-dollar100-billion-datacenter-project-for-an-ai-supercomputer

KathyReid, to stackoverflow
@KathyReid@aus.social avatar

I just issued a data deletion request to #StackOverflow to erase all of the associations between my name and the questions, answers and comments I have on the platform.

One of the key ways in which #RAG works to supplement #LLMs is based on proven associations. Higher ranked Stack Overflow members' answers will carry more weight in any #LLM that is produced.

By asking for my name to be disassociated from the textual data, it removes a semantic relationship that is helpful for determining which tokens of text to use in an #LLM.

If you sell out your user base without consultation, expect a backlash.

mcc, to random
@mcc@mastodon.social avatar

Hard to imagine a signal that a website is a rugpull more intense than banning users for trying to delete their own posts

https://www.tomshardware.com/tech-industry/artificial-intelligence/stack-overflow-bans-users-en-masse-for-rebelling-against-openai-partnership-users-banned-for-deleting-answers-to-prevent-them-being-used-to-train-chatgpt

Like just incredible "burning the future to power the present" energy here

chris,
@chris@strafpla.net avatar

@mcc So developers will stop sharing information on and future and friends will be forever stuck in the past, answering questions about historically relevant frameworks and languages.

ALTAnlp, to Korean
@ALTAnlp@sigmoid.social avatar

ICYMI: CALL FOR SPONSORS FOR Workshop

Does your company work in the space? Work with or technology?

We are now announcing our Call for for our upcoming at in beautiful .

This year, we’ve created budget-friendly packages starting from AU$500.

If you're interested please contact our Sponsorship Chair, Mr Charbel El-Khaissi at Charbel.El-Khaissi@anu.edu.au.

Boosts welcome ✅

noellemitchell, (edited ) to ai
@noellemitchell@mstdn.social avatar

The fact that everything I post on the Internet will most likely be scraped to train AI doesn't give me a lot of motivation to share stuff online 😅 1/

metin, to ai
@metin@graphics.social avatar

This is pretty cool. Curious what discoveries lie ahead…

𝘈𝘭𝘱𝘩𝘢𝘍𝘰𝘭𝘥 3 𝘱𝘳𝘦𝘥𝘪𝘤𝘵𝘴 𝘵𝘩𝘦 𝘴𝘵𝘳𝘶𝘤𝘵𝘶𝘳𝘦 𝘢𝘯𝘥 𝘪𝘯𝘵𝘦𝘳𝘢𝘤𝘵𝘪𝘰𝘯𝘴 𝘰𝘧 𝘢𝘭𝘭 𝘰𝘧 𝘭𝘪𝘧𝘦'𝘴 𝘮𝘰𝘭𝘦𝘤𝘶𝘭𝘦𝘴

https://blog.google/technology/ai/google-deepmind-isomorphic-alphafold-3-ai-model/

emill1984, to IT Polish
@emill1984@101010.pl avatar

Kiedys bylo "kazdy moze byc programista i zarabiac 15k miesiecznie"

Dzisiaj "kazdy moze byc prompt engineerem" xD Ciekawe kiedy to ebnie ;)

https://www.theverge.com/2024/5/8/24151847/microsoft-copilot-rewrite-prompt-feature-microsoft-365

SztucznaInteligencja

metin, (edited ) to blender
@metin@graphics.social avatar

Tried Leiapix's automatic depth algorithm on an old 3D-rendered image of mine.

Nice result out of the box, with only a few minor errors here and there.

https://www.leiapix.com

linuxgal, to llm
@linuxgal@techhub.social avatar
informapirata, to llm Italian
@informapirata@mastodon.uno avatar

Come funzionano gli , spiegato senza matematica

Da dove proviene l’apparente intelligenza di questi modelli. In questo articolo, cercherò di spiegare in termini semplici e senza utilizzare la matematica avanzata come funzionano i modelli di testo generativi, per aiutarti a pensarli come algoritmi informatici e non come magia.

@aitech

https://blog.miguelgrinberg.com/post/how-llms-work-explained-without-math

adelgado, to llm
@adelgado@eu.mastodon.green avatar

Would be nice to have a LLM that you can train locally with your organization documentation, to be able to have an interface to easily find that information buried in decades of documents
#LLM #MachineLearning #documentation #FOSS

huey, to Law
@huey@social.ketupat.me avatar

"When I was asked to beta test its AI research bot, I informed a major legal research provider that it worse than sucked. It was dangerous. Not only did it hallucinate... but it conflated almost all the critical distinctions that make law work. It failed to distinguish between jurisdictions, both states and state and federal, as well as majority, concurrences and dissents. To AI, it was all the same, words about law..."

https://www.lexblog.com/2024/05/03/all-rise-for-judge-ai/

#Law #LegalTech #LLM #AI

boilingsteam, to microsoft
@boilingsteam@mastodon.cloud avatar
janriemer, to LLMs
bananabob, to ai
@bananabob@mastodon.nz avatar

Can I take your order – and your data? The hidden reason retailers are replacing staff with AI bots

https://theconversation.com/can-i-take-your-order-and-your-data-the-hidden-reason-retailers-are-replacing-staff-with-ai-bots-229202

pseudocurious, to llm

@snoopy Je n'ai pas accès à ton billet pour la sur les . C'est normal ?

vitriolix, to OpenAI
@vitriolix@mastodon.social avatar

I wonder how many billions of dollars OpenAI is paying Stack Overflow for their db

"Stack Overflow signs deal with OpenAI to supply data to its models"

https://techcrunch.com/2024/05/06/stack-overflow-signs-deal-with-openai-to-supply-data-to-its-models/

#openai #ai #llm #chatgpt #stackoverflow

jemoka, to llm
@jemoka@maly.io avatar

🎉 new preprint day

Wrote some multi-hop reasoning work recently, formalizing inference as a

achieved results on game of 24 problem from tree of thougchts

https://arxiv.org/abs/2404.19055

SomeGadgetGuy, (edited ) to tech
@SomeGadgetGuy@techhub.social avatar

Replay Crew! We had a fun romp through tech headlines this week! https://somegadgetguy.com/b/44j
Jack Dorsey is no longer on the board of BlueSky. We're wrapping up the closing arguments in Google's anti-trust case. The Rabbit R1 is an app. Sony's marketing materials for the next XPERIA leak.

And we should probably chat about this next iPad thing-y...

  • All
  • Subscribed
  • Moderated
  • Favorites
  • normalnudes
  • InstantRegret
  • thenastyranch
  • mdbf
  • Durango
  • Youngstown
  • slotface
  • hgfsjryuu7
  • vwfavf
  • rosin
  • kavyap
  • osvaldo12
  • PowerRangers
  • DreamBathrooms
  • anitta
  • magazineikmin
  • khanakhh
  • GTA5RPClips
  • ethstaker
  • cubers
  • ngwrru68w68
  • tacticalgear
  • everett
  • tester
  • Leos
  • cisconetworking
  • modclub
  • provamag3
  • All magazines