metin, (edited ) to blender
@metin@graphics.social avatar

Tried Leiapix's automatic depth algorithm on an old 3D-rendered image of mine.

Nice result out of the box, with only a few minor errors here and there.

https://www.leiapix.com

chikim, to llm
@chikim@mastodon.social avatar

I created a multi-needle in a haystack test where a randomly selected secret sentence was split into pieces and scattered throughout the document with 7.5k tokens in random places. The task was to find these pieces and reconstruct the complete sentence with exact words, punctuation, capitalization, and sequence. After running 100 tests, llama3:8b-instruct-q8 achieved a 44% success rate, while llama3:70b-instruct-q8 achieved 100%! https://github.com/chigkim/haystack-test

pixelate, to accessibility
@pixelate@tweesecake.social avatar

Please boost for reach if this kind of stuff interests you. Will post more on this later.

Once upon a time, there was a cool emulator frontend called Retroarch. This emulator wasn't accessible until I and a few other gamers went to them and asked about adding accessibility. An amazing person known as BarryR made it happen. Now, if you turn on accessibility mode in settings, or pass the "--accessibility" (or something like that) flag on the command line, you get spoken menus, including the emulator's pause menu, good for saving states and such. Then, using PIL and other image processing Python utilities, running a server and hooking into Retroarch, the script allowed players to move around the map, battle, talk to NPC's, ETC. The only problem was, no one wanted to test it. The blind gaming community pretty much spoke, saying that we want new games. We want cool new, easy accessibility. So that's what we have no, follow the beacon or get sighted help in the case of diablo and such. It's sad, but meh. It's what we wanted I guess. No Zelda for us. So, this is about as far as he got:

To expand on what devinprater was saying: I am working on an accessibility pack/service for Final Fantasy 1 for the NES (this was what was shown in the latest RetroArch update). The idea is similar to how Pokemon Crystal access works, but it's using the RetroArch AI Service interface to do so.
Right now, the FF1 access service is mostly done, but I need more testers to try it out and give me feedback on how it's working. Right now, you can get up to the point where you get the ship, but there's no code to deal with how the ship moves, so that still needs to be done. Likewise with the airship later on.
The service works the latest version of RetroArch, on linux and mac, but not windows. This is due to how nvda reads out the text and until the next major update to nvda (which will have a feature to fix this), it'll have to wait. If you have those, I (or maybe devinprater) can help you set it up on mac/linux to test out. The package itself is available at: https://ztranslate.net/download/ff1_pac … zip?owner=

osi, to opensource
@osi@opensource.org avatar

Entering a new phase of the process: Validating the draft Open Source AI Definition. Which of these models comply with the principles of Open Source? Volunteer to analyze them!
https://discuss.opensource.org/t/definition-validation-seeking-volunteers/320

timbray, (edited ) to photography
@timbray@cosocial.ca avatar

Twenty years ago, worried about how Photoshop could be used to lie, I published a blog piece called “Photointegrity”. Yesterday I published another one with the same title: https://www.tbray.org/ongoing/When/202x/2024/04/29/Photointegrity

It considers “photointegrity” in the context of gen-AI imaging tools, like for example those recently announced by Adobe: https://petapixel.com/2024/04/23/adobe-introduces-one-of-its-most-significant-photoshop-updates-ever/ - which I find kind of terrifying,

Includes groovy pictures combining retro hardware with modern (non-generative) AI.

chikim, to llm
@chikim@mastodon.social avatar

VOLlama v0.1.0, an open-source, accessible chat client for OLlama
Unfortunately, many user interfaces for open source large language models are either inaccessible or annoying to use with screen readers, so I decided to make one for myself and others. Non screen reder users are welcome to use it as well.
I hope that ML UI libraries like Streamlit and Gradio will become more friendly with screen readers in the future, so making apps like this is not necessary!

https://chigkim.github.io/VOLlama/

chikim, to llm
@chikim@mastodon.social avatar

Wow, Private LLM locally runs Llama-3-8B on iOS. No idea how accessible with VoiceOver though. https://privatellm.app/en

cassidy, (edited ) to ai
@cassidy@blaede.family avatar

I really like the convention of using ✨ sparkle iconography as an “automagic” motif, e.g. to smart-adjust a photo or to automatically handle some setting. I hate that it has become the defacto iconography for generative AI. 🙁

metin, to ai
@metin@graphics.social avatar
chikim, to llm
@chikim@mastodon.social avatar

Tired of neutral responses from LLMs? Llama-3 seems great at following system prompts, so try this system prompt for an opinionated chatbot.
"You are a helpful, opinionated, decisive assistant. When asked a yes/no question, begin your respond with one word answer: yes or no. For open-ended or complex questions, adopt a firm stance. Justify your views with well-reasoned arguments, robust evidence, and succinct explanations, ensuring clarity and confidence in every response."

chikim, to llm
@chikim@mastodon.social avatar

Mark Zuckerberg on Llama 3: Apparently Meta stopped training Llama-3-70b before convergence and decided to move onto Llama-4. Meaning they could have kept training and made it smarter! Also llama3-70b multimodal as well as multilingual and bigger context window are coming. https://youtu.be/bc6uFV9CJGg

chikim, to llm
@chikim@mastodon.social avatar

Start saving money for that M4 Ultra with 500GB! Maybe this could be the first open source that could surpass GPT-4! AIatMeta: "Llama 3 8B & 70B models are just the beginning of what we’re working to release for Llama 3. Our largest models currently in the works are 400B+ parameters and while they’re still in active development, we’re excited about how this work is trending." https://twitter.com/AIatMeta/status/1780997414071181370

chikim, to llm
@chikim@mastodon.social avatar
cigitalgem, (edited ) to llm
@cigitalgem@sigmoid.social avatar

I am giving a Machine Learning Security talk in Bergen, Norway Monday 22.4. The scandanavian tour's last and final stop.

I will be discussing security risks and the new BIML results. Please join us.

Thanks for your help spreading the word.

https://www.lyyti.fi/reg/CDR-NO-22-04-2024

cigitalgem, to infosec
@cigitalgem@sigmoid.social avatar

Software Security Seminar in Stockholm TOMORROW 17.4

Please join me for an early morning breakfast seminar on (with some thrown in for good measure). Build security in.

Register here https://www.lyyti.fi/reg/CDR-SV-17-04-2024

Thank you in advance for passing this on to dev types you know in Sweden. Please boost for reach.

chikim, to ML
@chikim@mastodon.social avatar

Earlier today, Microsoft released new WizardLM-2 7b, 8x22b, 70b with great benchmark result, (of course, they say as good or almost same as GPT-4), but they removed weights on Huggingface, repo on Github, and their whitepaper. Someone on Reddit joked maybe they released GPT-4 by mistake! lol Quantized. weights from other people are still around on Huggingface!

chikim, to macos
@chikim@mastodon.social avatar

Cool tip for running LLMs on Apple Silicon! By default, MacOS allows GPU to use up to 2/3 of RAM on machines with <=36GB and 3/4 on machines with >36GB. I used the command sudo sysctl iogpu.wired_limit_mb=57344 to override and allocate 56GB/64GB for GPU. This allowed me to load all layers of larger models for a faster speed!

metin, to ai
@metin@graphics.social avatar

When generative AI is trained with AI-generated data, it becomes degenerat(iv)e AI.

datasciencejobsusa, to datascience
@datasciencejobsusa@mastodon.social avatar
pydatamadrid, to python Spanish
@pydatamadrid@masto.ai avatar

Ya está abierto el registro para nuestra reunión de abril: 🐲 LLMOps & ML para Drilling Performance y Python & Mazmorras, este mes en las oficinas de Repsol

https://www.meetup.com/pydata-madrid/events/300310880/

¡Nos vemos el jueves 18 a las 19:00! Y después, networking 🍻

chikim, to llm
@chikim@mastodon.social avatar

Thanks to all the recent large LLMs, "Apple is considering support for up to half a terabyte of RAM" for the highest-end m4 Mac configurations. I'm sure the price won't be cheap, but I bet it will be cheaper than getting 500GB in vram from NVidia. lol https://9to5mac.com/2024/04/11/apple-first-m4-mac-release-ai/

craigbrownphd, to technology
@craigbrownphd@mastodon.social avatar
datasciencejobsusa, to datascience
@datasciencejobsusa@mastodon.social avatar
metin, (edited ) to ai
@metin@graphics.social avatar

Whenever I see OpenAI's Sam Altman with his pseudo-innocent glance, he always reminds me of Carter Burke from Aliens (1986), who deceived the entire spaceship crew in favor of his corporation, with the aim of getting rich by weaponizing a newly discovered intelligent lifeform.

chikim, to llm
@chikim@mastodon.social avatar

Apparently Meta is planning to release two small varients of Llama-3 next week "as a precursor to the launch of the biggest version of Llama 3, expected this summer." Command-r-plus, mixtral 8x22b, Google CodeGemma... All of sudden companies are releasing LLMS like crazy! Where's Apple? Maybe In WWDC 2024? lol https://www.theinformation.com/articles/meta-platforms-to-launch-small-versions-of-llama-3-next-week

  • All
  • Subscribed
  • Moderated
  • Favorites
  • anitta
  • InstantRegret
  • thenastyranch
  • mdbf
  • khanakhh
  • Youngstown
  • slotface
  • hgfsjryuu7
  • ngwrru68w68
  • rosin
  • kavyap
  • Durango
  • PowerRangers
  • DreamBathrooms
  • Leos
  • magazineikmin
  • modclub
  • tacticalgear
  • osvaldo12
  • GTA5RPClips
  • everett
  • ethstaker
  • vwfavf
  • tester
  • normalnudes
  • cisconetworking
  • cubers
  • provamag3
  • All magazines