@chikim@mastodon.social avatar

chikim

@chikim@mastodon.social

Love music, technology, accessibility! Faculty at Berklee College of Music 👨🏻‍💻🎹🐕‍🦺

This profile is from a federated server and may be incomplete. Browse more on the original instance.

vick21, to accessibility
@vick21@mastodon.social avatar

How NVDA & OSARA are empowering blind people globally - Audio described Version: https://youtube.com/watch?v=N-y3yomLLSk&si=xiibf5ZxJzrlDnES

chikim,
@chikim@mastodon.social avatar

@vick21 I understand this is not a good way to measure by any mean, but regardless let's do some math! Globally there are 43M blind people , According to WHO. There are over 250k NVDA users in 175 countries, According to NVDA creators. Let's just say 300k. According to screen reader survey , NVDA counts for 37.7%. That means only 1.85% of blind people have access to screen reader. 0.3M/0.377%/43M*100 That's very sad! :( Let me know if I epically failed this math. lol

chikim,
@chikim@mastodon.social avatar

@miki @pixelate @vick21 Good point! Globally 82% of blind people are over 50. If you adjust the stat to look only under 50 years old, it's 11.33%. I think also screen reader survey includes screen readers on mobile devices, so it's not just computer.

chikim, to random
@chikim@mastodon.social avatar

I created samples for all 58 voices for xtts-v2. Hopefully it makes it easier for someone to choose a speaker. https://we.tl/t-9vWd1gO3EN

chikim, to apple
@chikim@mastodon.social avatar

Am I the only one who hates the giant escape key on newer Macbook Pro? It pushes all the function keys to the right, and it really throws off my muscle memory! Also Logitech thinks it's a great idea on MX Keys S. I just got older Logitech MX Mini for this reason. :( #apple #logitech #keyboard

chikim, to ai
@chikim@mastodon.social avatar

lol Cheery! The best name for ChatGPT 4O voice! Jeff Jarvis on TWIT podcast suggested. I agree, she's so over the top!

blindbargains, to random
@blindbargains@mastodon.social avatar

Be My Eyes Accessibility with GPT-4o https://www.youtube.com/watch?v=KwNUJ69RbwY

chikim,
@chikim@mastodon.social avatar

@FreakyFwoof @blindbargains Is gpt-4O on BeMyAI available now? Only for Beta users?

chikim, to llm
@chikim@mastodon.social avatar

Interesting, the Chat GPT desktop app for the Mac will be slowly rolling out to Plus subscribers starting today, but OpenAI "plans to release a Windows version of the desktop app later this year." Maybe the rumor, that Apple is cosing a deal with OpenAi for ChatGPT, is True... https://www.macrumors.com/2024/05/13/chatgpt-mac-app/

chikim, to llm
@chikim@mastodon.social avatar

If you missed it, check out the new GPT4O demo. https://www.youtube.com/watch?v=DQacCB9tDaw

chikim,
@chikim@mastodon.social avatar

@bryansmart lol I'm the one who posted about Google IO, and you replied to me. haha

chikim,
@chikim@mastodon.social avatar

@bryansmart FAlso funny. Google IO event is tomorrow. OpenAI intentionally revealed their model today. lol

chikim, to llm
@chikim@mastodon.social avatar

GPT4O combines audio, image, text. It can analyze actual audio. Also you can interrupt voice. It can pick up emotion from audio. You can also ask different speech with diffetrent style including singing! It can see the image in real time and chat on voice. For example solving equation in real time as you write on paper. This is amazing!

chikim, to accessibility
@chikim@mastodon.social avatar

I'm late to the party, but I found out I'm with the majority! lol According to the WebAIM Screen Reader User Survey , 68.2% (779 out of 1142) "respondents indicate that individuals should not describe what they look like during a virtual meeting or webinar" for blind and visually impaired participants. https://www.webaxe.org/webaim-screen-reader-user-survey-10/

chikim, to ai
@chikim@mastodon.social avatar

ElevenLabs joined the AI music generation. not available for public, but here's a demo clip. It's getting better and better! #AI #ML https://www.youtube.com/watch?v=m9DrkOrr3QM

chikim, to random
@chikim@mastodon.social avatar

Well Logic Pro on Ipad now has built in stem separation like Demucs. lol

macrumors, to random
@macrumors@mastodon.social avatar
chikim,
@chikim@mastodon.social avatar

@macrumors Hi, do you guys have MacRumorsLive on Mastodon?

bryansmart, to random

@chikim I've really been enjoying VOLlama. Nice work! Would be nice to be able to switch between OpenAI and local models without going in to API prefs. More accelerator keys for menu options would be good, too. Could maybe a blank line be inserted in the log between each entry? Last, can you trap key-down on the Control key to stop the system voice? I know it's a hobby project, so no idea how much time you have for any of that, but just throwing them out there.

chikim,
@chikim@mastodon.social avatar

@bryansmart the Alternative is press alt or option+up to go into the edit mode, and it'll paste only one message at a time into the prompt field.

chikim,
@chikim@mastodon.social avatar

@bryansmart Alt or option+up/down will let you edit the history context. It'll show one message at a time into the prompt field and lets you edit. If you want to just review it without editing it, just escape or alt/option+dwn all the way to the bottom when you're done reading.

chikim,
@chikim@mastodon.social avatar

@bryansmart Hmm, what kind of error? @vick21 also told me he got out of memory error during creating embedding.

chikim,
@chikim@mastodon.social avatar

@bryansmart @vick21 ah, you forgot to download the embedding model. ollama pull nomic-embed-text

chikim,
@chikim@mastodon.social avatar

@bryansmart @vick21 No problem. I totally get it. I hate writing manual/user guide/readme as well as reading them. lol

chikim,
@chikim@mastodon.social avatar

@vick21 @bryansmart That's actually great idea! VOLLlama has only few files, so I should try feeding everything and ask to write manual for it. hahahaha

chikim,
@chikim@mastodon.social avatar

@bryansmart Did you start your question with /q?

chikim,
@chikim@mastodon.social avatar

@bryansmart Also, you can't ask general questions like summarize that require reading the entire thing. The way it works it compares your question against documents through embedding, and retrieve and feed few chunks that might be relevant to answer your question.

chikim,
@chikim@mastodon.social avatar

@bryansmart If you want to see which chunks are fed into the model in order to answer your question, you can check the only checkbox in RAG settings. lol I don't think it's labeled in mac due to wx widget bug on Mac.

chikim,
@chikim@mastodon.social avatar

@bryansmart Embedding model is only used during indexing. The quality of the answer depends on the model you are chatting with. Because it will read the chunks in text and give you the answer. It really depends on if the LlamaIndex was able to retrieve the relevant chunks or not. You can increase number of chunks and chunk length, but you might end up feeding chunks that are not related to your question. Also there's threshold you can change to filter out below certain simularity score.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • megavids
  • kavyap
  • DreamBathrooms
  • everett
  • magazineikmin
  • InstantRegret
  • ngwrru68w68
  • Youngstown
  • mdbf
  • slotface
  • vwfavf
  • tacticalgear
  • thenastyranch
  • rosin
  • Leos
  • osvaldo12
  • ethstaker
  • GTA5RPClips
  • khanakhh
  • Durango
  • tester
  • normalnudes
  • cisconetworking
  • modclub
  • cubers
  • provamag3
  • anitta
  • JUstTest
  • All magazines