Hey, European musicians, what do you say about this nonsense? lol
Me: Could you explain what is H flat chord in harmony?
Gpt-4o: Certainly! In harmony, an H flat chord is typically known as a B flat chord in most musical contexts. The term "H" for B natural is used in some European countries, but "H flat" would be B flat in those regions as well.
VOLlama v0.1.4-beta.1: System Prompt manager; Import Awesome ChatGPT Prompts; Partial support for GPT-4O (Throws an error for token counter in some cases but just ignore for now); Able to attach entire document and feed for long context model. https://chigkim.github.io/VOLlama/
@simon I don't have Anthropic api, so it'd be hard to test and implement unfortunately. Do they have like pay as you go plan for API instead of monthly sub?
@jscholes@simon Q: Can I use the Claude API for individual use? A: No. Access to the API is subject to our Commercial Terms of Service and is not intended for individual use.
@serrebi@simon@jscholes Lookss like OpenRouter also uses OpenAI api end point, so you could try setting OPENAI_BASE_URL and OPENAI_API_KEY environment variable before running VOLlama. I haven't tried it.
@simon@serrebi@jscholes Don't bother. I just tried, and it doesn't work. I use Llamaindex, and apparently it's very specifically written for OpenAI. LlamaIndex has OpenAILike module for generic OpenAI API that's not OpenAI.com, but it doesn't seem to work. I need to look into it more.
BeMyEyes Privacy Policy 1/2: We record and store video streams and other images to enforce our Terms of Service, to promote and preserve safety, and to improve our Services and create new Services. We may provide recorded video streams or images to other organizations that are performing research or working to develop products and services that may assist blind and low-vision people or other members of the general public.
@JonathanMosen Sorry for specifically tagging you, but do you have any thought on the BeMyEyes privacy policy re images and videos you submit to the platform? It's too long to include in one post, but I pasted the relevant quotes in this thread. Should blind folks just say no such thing as free lunch and move on?
@twynn@JonathanMosen For paid customer, OpenAI has feature to opt out from getting your data used for training. Also if you use API, your data "do not become part of the training data unless you explicitly opt in." Your data is deleted within 30 days unless required for legal reasons, and is only accessible by authorized OpenAI employees, as well as specialized third-party contractors (that are subject to confidentiality and security obligations). https://www.maginative.com/article/openai-clarifies-its-data-privacy-practices-for-api-users/
@twynn@JonathanMosen Actually, if you have a free OpenAI account, you can turn off Improve the model for everyone in Settings > Data controls on this webpage. How much can we trust them? It's different story. lol https://chatgpt.com/#settings/DataControls
As I understand it, with all current LLMs, having a conversation involves feeding the model the entire conversation up to this point. That is, there is no memory: the prompt you feed it just gets longer and longer. So how does that work with something like GPT-4O which could be processing audio and/or video at a much faster rate? Surely the prompts must get very large very quickly with anything beyond a short interaction? Doesn't that mean the responses take longer and cost more as the conversation gets longer?
@jcsteh I don't know how it works, so everything is my speculation. lol Anyways, ChatGPT has memory feature. Possibly Retrieval-Augmented Generation? Also when you are about to reach the context limit, maybe they ask model to summarize previous context, and discard the detail and keep the important ones. Also maybe multimodal has longer context? For example, Google Gemini 1.5 Pro has 2 millions context length!