The “why would they make this” people don’t understand how important this type of research is. It’s important to show what’s possible so that we can be ready for it. There are many bad actors already pursuing similar tools if they don’t have them already. The worst case is being blindsided by something not seen before.
The rest of the budget kind of sucks but this part makes sense. If you’re making significant profits off of users in a country you should have to pay some of that back. All countries should have this.
Cohere’s command-r models are trained for exactly this type of task. The real struggle is finding a way to feed relevant sources into the model. There are plenty of projects that have attempted it but few can do more than pulling the first few search results.
I don’t think the term open-source can be applied to model weights. Even if you have the exact data, config, trainer and cluster it’s basically impossible to reproduce an exact model. Calling a model “open” sort of works but then there’s the distinction between open for research and open for commercial use. I think it’s kind of similar to the “free” software distinction. Maybe there’s some Latin word we could use.
Your best bet would probably be to get a used office PC to put the card in. You’ll likely have to replace the power supply and maybe swap the storage but with how much proper external enclosures go for the price might not be too different. Some frameworks don’t support direct GPU loading so make sure that you have more ram than vram.
An arm soc won’t work in most cases due to a lack of bandwidth and software support. The only board I know of that can do it is the rpi5 and that’s still mostly a poc.
In general I wouldn’t recomend a titan x unless you already have one because it’s been deprecated in cuda, so getting modern libraries to work will be a pain.
I really like the simplicity and formatting of stock pacman. It’s not super colorful but it’s fast and gives you all of the info you need. yay (or paru if you’re a hipster) is the icing on top.
This has more to do with how much chess data was fed into the model than any kind of reasoning ability. A 50M model can learn to play at 1500 elo with enough training: adamkarvonen.github.io/…/chess-world-models.html
The “AI PC” specification requires a minimum of 40TOPs of AI compute which is over double the 18TOPs in the current M3s. Direct comparison doesn’t really work though.
What really matters is how it’s made available for development. The Neural engine is basically a black box. It can’t be incorporated into any low level projects because it’s only made available through a high-level swift api. Intel by comparison seems to be targeting pytorch acceleration with their libraries.
This article is grossly overstating the findings of the paper. It’s true that bad generated data hurts model performance, but that’s true of bad human data as well. The paper used opt125M as their generator model, a very small research model with fairly low quality and often incoherent outputs. The higher quality generated data which makes up a majority of the generated text online is far less of an issue. The use of generated data to improve output consistency is a common practice for both text and image models.
It’s size makes it basically useless. It underperforms models even in it’s active weight class. It’s nice that it’s available but Grok-0 would have been far more interesting.
I feel like the whole Reddit AI deal is a trap. If any real judgment comes down about data use Reddit is an easy scapegoat. There was basically nothing stopping them from scraping the site for free.
Don’t buy a Chromebook for linux. While driver support usually isn’t an issue, the alternative keyboard layout is terrible for most applications. To even get access to all of the normal keys that many applications expect you need to configure multi-key shortcuts which varies in complexity based on your DE. In most cases it will also void your warranty because of the custom firmware requirement.
Tiny plastic shards found in human testicles, study says (www.cnn.com)
Critics question tech-heavy lineup of new Homeland Security AI safety board (arstechnica.com)
Steam will stop issuing refunds if you play two hours of a game before launch day (www.theverge.com)
Closing the early access loophole.
DuckDuckGo AI Chat (duckduckgo.com)
DDG is now offering free/private AI chat using several models.
Microsoft’s VASA-1 can deepfake a person with one photo and one audio track (arstechnica.com)
Zuckerberg says Meta's Llama 3 is really good but no chatbot is sophisticated enough to be an 'existential' threat — yet (www.businessinsider.com)
Canada to start taxing tech giants in 2024 despite U.S. complaints (www.bnnbloomberg.ca)
GPT-4 performance comparable with physicians on official medical board residency examinations. Model performance near or above official passing rate in all medical specialties tested (ai.nejm.org)
Instagram will blur nudes in messages sent to minors (www.theverge.com)
Guide on how to reverse a hate symbol that you can't remove (cdn.masto.host)
The tech industry can’t agree on what open-source AI means. That’s a problem. (www.technologyreview.com)
Meta’s AI image generator really struggles with the concept of interracial couples | CNN Business (www.cnn.com)
Meta’s AI image generator is coming under fire for its apparent struggles to create images of couples or friends from different racial backgrounds.
‘The machine did it coldly’: Israel used AI to identify 37,000 Hamas targets (www.theguardian.com)
Dock GPU to Laptop or to small SOC?
Afaik most LLMs run purely on the GPU, dont they?...
What is the most visually pleasing package manager (in terminal)?
MIT scientists have just figured out how to make the most popular AI image generators 30 times faster (www.livescience.com)
How well can LLMs solve chess puzzles? (github.com)
Each LLM is given the same 1000 chess puzzles to solve. See puzzles.csv. Benchmarked on Mar 25, 2024....
Mistral 7B v0.2 Base (released at SHACK15sf hackathon) (twitter.com)
GitHub: github.com/mistralai-sf24/hackathon...
Reddit is going public. Will its unruly user base revolt? (www.vox.com)
Microsoft’s first AI PCs are the Surface Pro 10 and Surface Laptop 6 for businesses (www.theverge.com)
Generative AI will eventually poison itself (www.xda-developers.com)
Grok-1 chatbot code released – open source or open Pandora's box? (www.theregister.com)
The FTC is probing Reddit’s AI licensing deals (www.engadget.com)
cross-posted from: slrpnk.net/post/7669534...
Fanless linux laptop
I’m looking for an Apple MacBook Air M2 alternative that could run Linux....