>>> Do you happen to know what your context window length is?
Llama: I'm an AI model, and I don't have a fixed "context window" in the classical sense. My training data consists of a massive corpus of text, which I use to generate responses.
#AI#GenerativeAI#LLMs#OpenSource#Microsoft#WizardLM2: "Last week, Microsoft researchers released WizardLM 2, which it claimed is one of the most powerful open source large language models to date. Then it deleted the model from the internet a few hours later because, as The Information reported, it “accidentally missed” required “toxicity testing” before it was released.
However, as first spotted by Memetica, in the short hours before it was taken down, several people downloaded the model and reuploaded it to Github and Hugging Face, meaning that the model Microsoft thought was not ready for public consumption and had to be taken offline, has already spread far and wide, and now effectively can never be removed from the internet.
#ML#AI#GenerativeAI#LLMs#FoundationModels#PoliticalEconomy: "A recent innovation in the field of machine learning has been the creation of very large pre-trained models, also referred to as ‘foundation models’, that draw on much larger and broader sets of data than typical deep learning systems and can be applied to a wide variety of tasks. Underpinning text-based systems such as OpenAI's ChatGPT and image generators such as Midjourney, these models have received extraordinary amounts of public attention, in part due to their reliance on prompting as the main technique to direct and apply them. This paper thus uses prompting as an entry point into the critical study of foundation models and their implications. The paper proceeds as follows: In the first section, we introduce foundation models in more detail, outline some of the main critiques, and present our general approach. We then discuss prompting as an algorithmic technique, show how it makes foundation models programmable, and explain how it enables different audiences to use these models as (computational) platforms. In the third section, we link the material properties of the technologies under scrutiny to questions of political economy, discussing, in turn, deep user interactions, reordered cost structures, and centralization and lock-in. We conclude by arguing that foundation models and prompting further strengthen Big Tech's dominance over the field of computing and, through their broad applicability, many other economic sectors, challenging our capacities for critical appraisal and regulatory response." https://journals.sagepub.com/doi/full/10.1177/20539517241247839
Like words, molecular sequences in biological components are tokens that can be manipulated by #LLMs:
“Here, using large language models (LLMs) trained on biological diversity at scale, we demonstrate the first successful precision editing of the human genome with a programmable gene editor designed with AI.”
@hrheingold Let's hope that works out better than the "novel materials" model, or at least that someone checks the sums before going anywhere near sn actual human patient, eh?
⚠️ @forrestbrazeal on the inside threat to OSS
🍴Vicki Boykis says Redis is forked
👻 @johnonolan says Ghost is federating
🦙 Meta Engineering announces Llama 3
❓ @eieio's questions to ask when you don't want to work
🎙 hosted by @jerod
Lots of people who work in #AI have, in their head, an idea about what sort of interaction with an #LLMmight give them pause. The thing that might make them start to suspect that something interesting is happening.
Here's mine:
User: Tell me a cat joke.
LLM: Why did the cat join a band? He wanted to be a purr-cussionist.
I was listing something on eBay, and they encourage starting with an existing listing—presumably to increase the amount of detail and decrease the amount of work.
When I selected the same model, I got a default description that was extremely robotic and wordy while just repeating the spec sheet. I thought it sounded LLM-generated; sure enough when I went to edit it, there is a big shiny “write with AI” button.
If you have documents with the answers you're looking for, why not search the documents directly? Why are you embedding the documents then using #RAG (Retrieval Augmenter Generation) to make a large language model give you answers? An LLM generates text, it doesn't search a DB to give you results. So just search the damn DB directly, we already have great search algorithms with O(1) retrieval speeds! #LLMs are so stupid.
@savvykenya Most jobs are fine with mediocre answers. If a new employee needs to know something and doesn’t know where to look, that’s a pretty good use-case for RAG.
Our upcoming “Build Your First #Python Package!” workshop has a section on how #LLMs like #ChatGPT can be used to support tasks such as documenting and formatting your code to improve usability and maintainability. We’ll also considering the ethical and logistical challenges, pitfalls and concerns associated with using #AI based tools in software development.
“But this doesn’t save any time!” 3 useful questions when trying #LLMs:
Is there another way to get results I want? Don't give up right away.
Does AI make this task less or more annoying? Sometimes supervising drudge work feels better even if it's not faster; other times you'd still rather do it yourself.
Are results likely to improve as LLMs get better? If so, add a calendar reminder to try again in a few months. Or, keep a list of things you want to re-try post GPT-5 class models. #GenAI
@smach i’ve actually found it useful to be vague at times. just give it a general direction and see what happens. a lot of times too much direction yields worse results
Here is an example of how bad #LLMs are with math. I asked about velocity in the context of Agile process. The answer?
“Sure! Let's say that an Agile development team has completed four iterations, each lasting two weeks. In the first iteration, they delivered 12 user stories; in the second, they delivered 10; in the third, they delivered 9; and in the fourth, they delivered 8. The total number of user stories completed by the end of the fourth iteration is 49 (12 + 10 + 9 + 8)”.
I would much rather get a "no results" when I'm looking for medical interactions than an #LLM helpfully telling me "Here's some bullshit you don't know enough to know is horribly wrong"
Even something as innocent as acetaminophen can destroy your liver if you overdose on it.