nsaphra, to random
@nsaphra@sigmoid.social avatar

New , led by Angelica Chen! We break the steepest MLM training loss drop into 2 phase changes: first in internal grammatical structure, then external capabilities. Big implications for emergence, simplicity bias, and interpretability! https://arxiv.org/abs/2309.07311

ramikrispin, to ChatGPT
@ramikrispin@mstdn.social avatar

A new crash course on Vector Embeddings from freeCodeCamp. The course, by Ania Kubow, focuses on applied applications of vector embeddings using GPT4 👇🏼

https://www.youtube.com/watch?v=yfHHvmaMkcA

ZachWeinersmith, to comics
@ZachWeinersmith@mastodon.social avatar
doboprobodyne, (edited )
@doboprobodyne@mathstodon.xyz avatar

@ZachWeinersmith

Re.: http://smbc-comics.com/comic/ancient-times

I beg your pardon, but my websearch-fu was no use in finding the origin of the words in your caption: "... as if summer storm had rainbowed the world, yet passed over your home as you dwelt in twilight and sorrow".

I thought it a particularly beautiful way to describe that peculiar sadness that is lifted from those around an individual by a global change in circumstance, but not from the individual, by whom a far more disabling insult had been received at or following the time of sadness.

May I ask, is it original Weinersmith?

ramikrispin, to llm
@ramikrispin@mstdn.social avatar

(1/2) New LLM model - DeciLM 6B-Instruct 🚀🚀🚀

Deci AI released today a new LLM - DeciLM 6B-Instruct. The DeciLM is an auto-regressive language model, and it was built by LoRA fine-tuning using a subset of the OpenOrca dataset. The model has 5.7 billion parameters, using a context window of 4096 tokens. According to the company, the model runs 15 times faster than Llama 2 7B while maintaining comparable quality.

https://huggingface.co/Deci/DeciLM-6b-instruct

mc, to ML

Tomorrow (9/11, 3 4pm GMT) Nouha Dziri from AI2 will present on her paper:

Faith and Fate: Limits of Transformers on Compositionality

at the Formal Languages and Neural Networks (FLaNN) Seminar!

Paper: https://arxiv.org/abs/2305.18654

Seminar Info: https://flann.super.site/

EDIT: I had the wrong time, it starts in 1 hour!

smtibor, to ArtificialIntelligence
@smtibor@fosstodon.org avatar

Just published the most recent episode of Teaching Python with @ines talking about NLP in education.

https://share.fireside.fm/episode/UIYXtbeL+uZjtzRf5

p.s. I need to get faster at publishing the podcast episode after the live stream.

ttpphd, to ArtificialIntelligence
@ttpphd@mastodon.social avatar

The author of package SpaCy has some educational thoughts about LLMs and why they are not the end point of NLP (at least not yet).

https://news.ycombinator.com/item?id=37443921

bwaber, to random
@bwaber@hci.social avatar

Did I mention it was hot? Running after my child while they sped along on a hoverboard wasn't exactly pleasant, but at least I was able to listen to some talks for my ! (1/11)

bwaber,
@bwaber@hci.social avatar

First was a nice talk by Boago Okgetheng on an system for Setswana and a panel on the startup journey of Amathambo AI with Ian Omung'a, Kira Düsterwald, and Sicelukwanda Zwane at . This impressive, inspiring work to learn about https://www.youtube.com/watch?v=lBOO7iJPADA (2/11)

bwaber, to random
@bwaber@hci.social avatar

It was a bit hot in Boston today (even the 🐢thought it was better to be inside), but I was still able to go for a shorter run and listen to talks for my ! (1/11)

bwaber,
@bwaber@hci.social avatar

Next was a nice group of short talks on various startups/projects at . I particularly liked the talks by Asmelash Teka Hadgu and Paul Azunre with targeted approaches https://www.youtube.com/watch?v=F21GuEZ8EoY (6/11)

ocramz, to ArtificialIntelligence
@ocramz@sigmoid.social avatar

Of the 3 papers I co-reviewed for , 2 used chatgpt for generation or evaluation.

Does anyone see a problem with littering our research with proprietary, black-box baselines?

BenjaminHan, to generativeAI
@BenjaminHan@sigmoid.social avatar

1/ In this age of LLMs and generative AI, do we still need knowledge graphs (KGs) as a way to collect and organize domain and world knowledge, or should we just switch to language models and rely on their abilities to absorb knowledge from massive training datasets?

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

2/ An early paper in 2019 [1] posited that compared to , it is easier for language models to adapt to new data without human supervision, and they allow users to query about an open class of relations without much restriction. To measure the knowledge encoding capability, the authors construct the LAMA (Language Model Analysis) probe where facts are turned into cloze statements and language models are asked to predict the masked words (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

3/ The result shows that even without specialized training, language models such as BERT-large can already retrieve decent amount of facts from their weights (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

4/ But is that all? A recent paper revisits this question and offers a different take [2]. The authors believe just testing isolated fact retrieval is not sufficient to demonstrate the power of KGs.

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

5/ Instead, they focus on more intricate topological and semantic attributes of facts, and propose 9 benchmarks testing modern LLMs’ capability in retrieving facts with the following attributes: symmetry, asymmetry, hierarchy, bidirectionality, compositionality, paths, entity-centricity, bias and ambiguity (screenshots).

image/png

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

6/ In each benchmark, instead of asking LLMs to retrieve masked words from a cloze statement, it also asks the LLMs to retrieve all of the implied facts and compute scores accordingly (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

7/ Their result shows that even achieves only 23.7% hit@1 on average, even when it scores up to 50% precision@1 using the earlier proposed LAMA benchmark (screenshot). Interestingly, smaller models like BERT can outperform GPT4 on bidirectional, compositional, and ambiguity benchmarks, indicating bigger is not necessarily better.

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

8/ There are surely other benefits of using KGs to collect and organize knowledge. They do not require costly retraining to update, therefore can be updated more frequently to remove obsolete or incorrect facts. They allow more trackable reasoning and can offer better explanations. They make fact editing more straightforward and accountable (think of GDPR) compared to model editing [3].

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

9/ But LLMs can certainly help in bringing in domain-specific or commonsense knowledge in a data-driven way. In conclusion: why not both [4]? :-)

BenjaminHan, to ai
@BenjaminHan@sigmoid.social avatar

Dog Lenat, founder of , passed away earlier this week. From Professor Ken Forbus:

"People in AI often don't give the Cyc project the respect it deserves. Whether or not you agree with an approach, understanding what has happened in different lines of work is important. The Cyc project was the first demonstration that symbolic representations and reasoning could scale to capture significant portions of commonsense…”

https://www.linkedin.com/posts/forbus_ai-knowledgegraphs-krr-activity-7103445990954700800-qcd-

ramikrispin, to llm
@ramikrispin@mstdn.social avatar

The Ask the SQL DB App 🦜🔗 is a cool Streamlit application made by
Harrison Chase and it is based on LangChain and LLM. This app translates the user questions into SQL queries 👇🏼

https://sql-langchain.streamlit.app

Code available here ➡️: https://github.com/hwchase17/sql-qa

DBuschek, to ai

Looking for participants: We currently run a survey on using ChatGPT for scientific writing. You'll revise one of your abstracts using ChatGPT and answer a few questions (ca. 10 mins.) https://tummgmt.eu.qualtrics.com/jfe/form/SV_1H2inMBL0ijC2ZU (can be switched to English or German at the top right)

BenjaminHan, to LLMs
@BenjaminHan@sigmoid.social avatar

1/ How robust and reliable is the code generated by , especially for real-world software development? A recent work [2] constructed a new benchmark based on [1] to evaluate if the generated code uses API correctly. Four popular -- .5, , , and -- are tested, and under zero-shot scored 62.09% misuse rate. Even with one-shot relevant examples the misuse rate of is 49.17%.

image/png
image/png
image/png

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

2/ Since users of with particular APIs are usually relatively inexperienced in the said APIs, these inaccuracies may have grave consequences to the robustness and reliability of the resulting software.

(How would fare?)

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

3/ REFERENCES

[1] Tianyi Zhang, Ganesha Upadhyaya, Anastasia Reinhardt, Hridesh Rajan, and Miryung Kim. 2018. Are code examples on an online Q&A forum reliable? a study of API misuse on stack overflow. In Proceedings of the 40th International Conference on Software Engineering, pages 886–896, Gothenburg, Sweden. Association for Computing Machinery. http://dx.doi.org/10.1145/3180155.3180260

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

4/end

[2] Li Zhong and Zilong Wang. 2023. A Study on Robustness and Reliability of Large Language Model Code Generation. http://arxiv.org/abs/2308.10335

  • All
  • Subscribed
  • Moderated
  • Favorites
  • provamag3
  • ngwrru68w68
  • InstantRegret
  • ethstaker
  • magazineikmin
  • osvaldo12
  • rosin
  • mdbf
  • Youngstown
  • khanakhh
  • slotface
  • Durango
  • kavyap
  • DreamBathrooms
  • JUstTest
  • thenastyranch
  • tacticalgear
  • cisconetworking
  • GTA5RPClips
  • modclub
  • cubers
  • normalnudes
  • everett
  • tester
  • megavids
  • Leos
  • anitta
  • lostlight
  • All magazines