BenjaminHan,
@BenjaminHan@sigmoid.social avatar

1/ In this age of LLMs and generative AI, do we still need knowledge graphs (KGs) as a way to collect and organize domain and world knowledge, or should we just switch to language models and rely on their abilities to absorb knowledge from massive training datasets?

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

2/ An early paper in 2019 [1] posited that compared to , it is easier for language models to adapt to new data without human supervision, and they allow users to query about an open class of relations without much restriction. To measure the knowledge encoding capability, the authors construct the LAMA (Language Model Analysis) probe where facts are turned into cloze statements and language models are asked to predict the masked words (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

3/ The result shows that even without specialized training, language models such as BERT-large can already retrieve decent amount of facts from their weights (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

4/ But is that all? A recent paper revisits this question and offers a different take [2]. The authors believe just testing isolated fact retrieval is not sufficient to demonstrate the power of KGs.

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

5/ Instead, they focus on more intricate topological and semantic attributes of facts, and propose 9 benchmarks testing modern LLMs’ capability in retrieving facts with the following attributes: symmetry, asymmetry, hierarchy, bidirectionality, compositionality, paths, entity-centricity, bias and ambiguity (screenshots).

image/png

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

6/ In each benchmark, instead of asking LLMs to retrieve masked words from a cloze statement, it also asks the LLMs to retrieve all of the implied facts and compute scores accordingly (screenshot).

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

7/ Their result shows that even achieves only 23.7% hit@1 on average, even when it scores up to 50% precision@1 using the earlier proposed LAMA benchmark (screenshot). Interestingly, smaller models like BERT can outperform GPT4 on bidirectional, compositional, and ambiguity benchmarks, indicating bigger is not necessarily better.

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

8/ There are surely other benefits of using KGs to collect and organize knowledge. They do not require costly retraining to update, therefore can be updated more frequently to remove obsolete or incorrect facts. They allow more trackable reasoning and can offer better explanations. They make fact editing more straightforward and accountable (think of GDPR) compared to model editing [3].

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

9/ But LLMs can certainly help in bringing in domain-specific or commonsense knowledge in a data-driven way. In conclusion: why not both [4]? :-)

tanepiper,
@tanepiper@tane.codes avatar

@BenjaminHan Exactly. Thanks for the references, hopefully next at the next we'll have a lot to talk about on this very topic (with LLMs) - but already the project I'm lead on, the work in our KG is vital for creating a strong semantic vocabulary we can use to connect and structure content.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • generativeAI
  • DreamBathrooms
  • InstantRegret
  • thenastyranch
  • magazineikmin
  • GTA5RPClips
  • rosin
  • osvaldo12
  • tacticalgear
  • Youngstown
  • slotface
  • khanakhh
  • kavyap
  • ngwrru68w68
  • modclub
  • Leos
  • everett
  • provamag3
  • cubers
  • cisconetworking
  • ethstaker
  • Durango
  • mdbf
  • anitta
  • megavids
  • normalnudes
  • tester
  • JUstTest
  • lostlight
  • All magazines