Skip to main content
Uncategorized

Evaluating the Knowledge Base Completion Potential of GPT

October 23, 2023

Blerta Veseli1, Simon Razniewski2, Jan-Christoph Kalo3, Gerhard Weikum1

Max Planck Institute for Informatics1, Bosch Center for AI2, University of Amsterdam3

{weikum,bveseli}@mpi-inf.mpg.de simon.razniewski@de.bosch.com j.c.kalo@uva.nl

Abstract

Structured knowledge bases (KBs) are an asset for search engines and other applications, but are inevitably incomplete. Language models (LMs) have been proposed for unsupervised knowledge base completion (KBC), yet, their ability to do this at scale and with high accuracy remains an open question. Prior experimental studies mostly fall short because they only evaluate on popular subjects, or sample already existing facts from KBs. In this work, we perform a careful evaluation of GPT’s potential to complete the largest public KB: Wikidata. We find that, despite their size and capabilities, models like GPT-3, ChatGPT and GPT-4 do not achieve fully convincing results on this task. Nonetheless, they provide solid improvements over earlier approaches with smaller LMs. In particular, we show that, with proper thresholding, GPT-3 enables to extend Wikidata by 27M facts at 90% precision.

1        Introduction

Structured knowledge bases (KBs) like Wiki- data (Vrandecˇic´ and Krötzsch, 2014), DBpedia (Auer et al., 2007), and Yago (Suchanek et al., 2007) are employed in many knowledge-centric applications like search, question answering and dialogue. Constructing and completing these KBs at high quality and scale is a long-standing re- search challenge, and multiple benchmarks exist, e.g., FB15k (Bordes et al., 2013), CoDEx (Safavi and Koutra, 2020), and LM-KBC22 (Singhania et al., 2022). Text-extraction, knowledge graph embeddings, and LM-based knowledge extraction have continuously moved scores upwards on these tasks, and leaderboard portals like Paperswithcode1 provide evidence for that.

Recently, LMs have been purported as a promising source of structured knowledge. Starting from the seminal LAMA paper (Petroni et al., 2019), a throve of works have explored how to better probe, train, or fine-tune these LMs (Liu et al., 2022).

Nonetheless, we observe a certain divide between these late-breaking investigations, and practical KB completion. While recent LM-based approaches often focus on simple methodologies that produce fast results, practical KBC so far is a highly precision-oriented, extremely laborious process, involving a very high degree of manual labour, either for manually creating statements (Vrandecˇic´ and Krötzsch, 2014), or for building comprehensive scraping, cleaning, validation, and normalization pipelines (Auer et al., 2007; Suchanek et al., 2007). For example, part of Yago’s success stems from its validated >95% accuracy, and according to (Weikum et al., 2021), the Google Knowledge Vault was not deployed into production partly be- cause it did not achieve 99% accuracy. Yet, many previous LM analyses balance precision and recall or report precision/hits@k values, implicitly tuning systems towards balanced recall scores resulting in impractical precision. It is also important to keep in mind the scale of KBs: Wikidata currently contains around 100 million entities and 1.2B statements. The cost of producing such KBs is massive. An estimate from 2018 sets the cost per statement at 2 $ for manually curated statement, and 1 ct for automatically extracted ones (Paulheim, 2018). Thus, even small additions in relative terms might correspond to massive gains in absolute numbers. For example, even by the lower estimate of 1 ct/statement, adding one statement to just 1% of Wikidata humans would come at a cost of 100,000 $.

In this paper, we conduct a systematic analysis of the KB completion potential of GPT, where we focus on high precision. We evaluate by employing (i) a recent KB completion benchmark, WD- KNOWN, (Veseli et al., 2023), which randomly samples facts from Wikidata and (ii) by a manual evaluation of subject-relation pairs without object values. Our main results are:

  1. For the long-tail entities of WD-KNOWN, GPT models perform considerably worse than what less demanding benchmarks like LAMA (Petroni et al., 2019) have indicated. Nonetheless, we can achieve solid results for language- related, socio-demographic relations (e.g., nativeLanguage).
  2. Despite their fame and size, out of the box, the GPT models, including GPT-4, do not pro- duce statements of a high enough accuracy as typically required for KB completion.
  3. With simple thresholding, for the first time, we obtain a method that can extend the Wikidata KB at extremely high quality (>90% precision), at the scale of millions of statements. Based on our analysis of 41 common relations, we would be able to add a total of 27M high- accuracy statements.

2        Background and Related Work

KB construction KB construction has a considerable history. One prominent approach is by hu- man curation, as done e.g., in the seminal Cyc project (Lenat, 1995), and this is also the back-bone of today’s most prominent public KB, Wiki- data (Vrandecˇic´ and Krötzsch, 2014). Another popular paradigm is the extraction from semi- structured resources, as pursued in Yago and DBpedia (Suchanek et al., 2007; Auer et al., 2007). Ex- traction from free text has also been explored (e.g., NELL (Carlson et al., 2010)). A popular paradigm has been embedding-based link prediction, e.g., via tensor factorization like Rescal (Nickel et al., 2011), and KG embeddings like TransE (Bordes et al., 2013).

An inherent design decision in KBC is the P/R trade-off – academic projects are often open to trade these freely (e.g., via F-1 scores), yet production environments are often critically concerned with precision, e.g., Wikidata generally discouraging statistical inferences, and industrial players likely use to a considerable degree human editing and verification (Weikum et al., 2021).

For example in all of Rescal, TransE, and LAMA, the main results focus on metrics like hits@k, MRR, or AUC, which provide no bounds on precision.

LMs for KB construction Knowledge extraction from LMs provides fresh hope for the synergy of automated approaches and high-precision curated KBs. It provides remarkably straightforward access to very large text corpora: The basic idea by (Petroni et al., 2019) is to just de- fine one template per relation, then query the LM with subject-instantiated versions, and retain its top prediction(s). A range of follow-up works appeared, focusing, e.g., on investigating entities, improving updates, exploring storage limits, incorporating unique entity identifiers, and others (Shin et al., 2020; Poerner et al., 2020; Cao et al., 2021; Roberts et al., 2020; Heinzerling and Inui, 2021; Petroni et al., 2020; Elazar et al., 2021; Razniewski et al., 2021; Cohen et al., 2023; Sun et al., 2023). Nonetheless, we observe the same gaps as above: The high-precision area, and completion of already existing resources, are not well investigated.

Several works have analyzed the potential of larger LMs, specifically GPT-3 and GPT-4,. They investigate few-shot prompting for extracting factual knowledge for KBC (Alivanistos et al., 2023) or for making the factual knowledge in a LM more explicit (Cohen et al., 2023). These models can aid in building a knowledge base on Wikidata or improving the interpretability of LMs. Despite the variance in the precision of extracted facts from GPT-3, it can peak at over 90% for some relations. Recently, GPT-4’s capabilities for KBC and reasoning were examined (Zhu et al., 2023). This research compared GPT-3, ChatGPT, and GPT-4 on information extraction tasks, KBC, and KG- based question answering. However, these studies focus on popular statements from existing KBs, neglecting the challenge of introducing genuinely new knowledge in the long tail.

In (Veseli et al., 2023), we analyzed to which degree BERT can complete the Wikidata KB, i.e., provide novel statements. Together with the focus on high precision, this is also the main difference of the present work to the works cited above, which evaluate on knowledge already existing in the KB, and do not estimate how much they could add.

3        Analysis Method

Dataset We consider the 41 relations from the LAMA paper (Petroni et al., 2019). For automated evaluation and threshold finding, we employ the WD-KNOWN dataset (Veseli et al., 2023). Unlike other KBC datasets, this one contains truly long-

    Relation GPT-3                    ChatGPT GPT-4
text-davinci-003         gpt-3.5-turbo
PRF 1PRF 1PRF 1
writtenIn0.620.590.60.910.780.840.370.850.52
ownedBy0.410.30.350.60.440.510.170.610.27
nativeLanguage0.850.850.850.690.860.770.530.880.66
LanguageOfFilm0.780.630.70.580.520.550.480.650.55
hasCapital0.770.480.590.770.440.560.490.860.62
officialLanguage0.620.60.610.670.640.650.270.730.39
foundedIn0.270.530.360.40.380.390.140.580.23
playsInstrument0.250.360.30.160.180.170.130.60.21
partOf0.050.10.060.170.10.130.060.360.1
citizenOf0.720.620.670.670.60.630.470.680.56
spokenLanguage0.480.620.540.540.760.630.370.840.51
playerPosition0.40.40.40.180.240.210.230.70.35
inContinent0.620.560.590.610.60.60.40.660.5
namedAfter0.50.490.490.530.440.480.120.360.18
hostCountry0.770.50.610.750.480.590.40.550.46
musicLabel0.290.310.30.160.180.170.080.480.14
hasReligion0.440.360.40.470.380.420.160.390.23
developedBy0.430.70.530.450.50.470.110.60.19
countryOfJurisdiction0.380.240.290.520.220.40.420.330.37
subclassOf0.210.40.280.730.850.790.160.620.25
diplomaticRelation0.360.620.460.70.170.270.410.320.36
CountryOfOrigin0.60.340.430.480.310.380.20.390.26
Macro-Average0.490.480.470.530.460.480.280.590.36

Table 1: Automated evaluation in the retain-all setting: GPT-3 (text-davinci-003 with 175B parameters), GPT-4 (#parameters unknown) and ChatGPT (gpt-3.5-turbo with #parameters unknown ) on 1000 samples/relation from WD-Known.

tail entities, by randomly sampling from Wikidata, a total of 4 million statements for 3 million sub- jects in 41 relations (Petroni et al., 2019). Besides this dataset for automated evaluation, for the main results, we use manual evaluation on Wikidata entities that do not yet have the relations of interest. For this purpose, for each relation, we manually define a set of relevant subject types (e.g., software for developedBy), that allows us to query for subjects that miss a property.

Evaluation protocol In the automated setting, we first use a retain-all setting, where we evaluate the most prominent GPT models (GPT-3 text- davinci-003, GPT-4, and ChatGPT gpt-3.5-turbo) by precision, recall, and F1. Table 1 shows that none of the GPT models could achieve precision of

>90%. In a second step, the precision-thresholding setting, we therefore sort predictions by confidence and evaluate by recall at precision 95% and 90% (R@P95 and R@P90). To do so, we sort the predictions for all subjects in a relation by the model’s probability on the first generated token2, then compute the precision at each point of this list, and return the maximal fraction of the list covered while maintaining precision greater than the desired value. We threshold only GPT-3, because only GPT-3’s token probabilities are directly accessable in the API, and because the chat-aligned models do not outperform it in the retain-all setting. Approaches to estimate probabilities post-hoc can be found in (Xiong et al., 2023).

Since automated evaluations are only possible for statements already in the KB, in a second step, we let human annotators evaluate the correctness of 800 samples of novel (out-of-KB) high-accuracy predictions. We hereby use a relation-specific threshold determined from the automated 75%-95% precision range. MTurk annotators could use Web search to verify the correct- ness of our predictions on a 5-point Likert scale (correct/likely/unknown/implausible/false). We counted predictions that were rated as correct or likely as true predictions, and all others as false.

Prompting setup To query the GPT models, we utilize instruction-free prompts listed in the appendix. Specifically for GPT-3, we follow the prompt setup of (Cohen et al., 2023), which is based on an instruction-free prompt entirely consisting of 8 randomly sampled and manually checked examples. In the default setting, all example subjects have at least one object. Since none of the GPT models achieved precision >90% and we can only threshold GPT-3 for high precision, we focus on the largest GPT-3 model (text-davinci-003) in the following. We experimented with three variations for prompting this model:

  1. Examples w/o answer: Following (Cohenet al., 2023), in this variant, we manually se- lected 50% few-shot examples, where GPT-3 did not know the correct answer, to teach the model to output “Don’t know”. This is sup- posed to make the model more conservative in cases of uncertainty.
  2. Textual context augmentation: Following (Petroni et al., 2020), we test whether adding textual context improves model performance. We hereby employ Google Web Search, with the subject and relation of interest as search query. The top 1 result snippet is then included as context to the prompt.
  3. #few-shot examples: A standard parameter in prompting is the number of few-shot examples. They have a huge impact on monetary costs. We vary this number between 1 and 12.
  Relation#current stmts. in Wikidata#subj. w/ missing stmts.fraction for which GPT-3 can give high-confidence prediction  #addable stmts.manual accuracyrelative growth
foundedIn43,254225,5789%20,30292%43%
citizenOf4,206,6844,616,6015%230,83082%5%
countryOfJurisdiction901,06624,96676%18,97488%2%
namedAfter340,234477,84522%105,12564%20%
inContinent71,101889,13462%551,26388%682%
ownedBy449,140416,4376%24,98624%1%
hostCountry14,275,59635,21453%18,66388%0%
spokenLanguage2,148,7757,134,54357%4,066,68992%174%
writtenIn14,140,45324,990,16173%18,242,81792%119%
officialLanguage19,6786,77642%2,846100%14%
developedBy42,37929,3496%1,76194%4%
CountryOfOrigin1,296,038135,19649%66,24630%2%
hasCapital111,17197311%10714%0%
LanguageOfFilm337,68270,66924%16,96182%4%
nativeLanguage264,7787,871,08549%3,856,83182%1195%
sharesBorders6,94622214%3172%0%
Overall38,654,97546,924,74966%27,224,43290 %70%

Table 2: Manual evaluation: Wikidata KB completion potential of GPT-3 text-davinci-003 with precision-oriented thresholding.

4        Results and Discussion

Can GPT models complete Wikidata at precision AND scale? In Table 1 we already showed that without thresholding, none of the GPT models can achieve sufficient precision. Table 2 shows our main results when using precision-oriented thresholding, on the 16 best-performing relations. The fourth column shows the percentage of subjects for which we obtained high-confidence predictions, the fifth how these translates into absolute statement numbers, and the sixth shows the percentages that were manually verified as correct (sampled). In the last column, we show how this number relates to the current size of the relation.

We find that manual precision surpasses 90% for 5 relations, and 80% for 11. Notably, the best- performing relations are mostly related to socio- demographic properties (languages, citizenship).

In absolute terms, we find a massive number of high-accuracy statements that could be added to the writtenIn relation (18M), followed by spokenLanguage and nativeLanguage (4M each). In relative terms, the additions could increase the existing relations by up to 1200%, though there is a surprising divergence (4 relations over 100%, 11 relations below 20%).

Does GPT provide a quantum leap? Generating millions of novel high-precision facts is a significant achievement, though the manually verified precision is still below what industrial KBs aim for. The wide variance in relative gains also shows that GPT only shines in selected areas. In line with previous results (Veseli et al., 2023), we find that GPT can do well on relations that exhibit high surface correlations (person names often give away their nationality), otherwise the task remains hard.

In Table 3 we report the automated evaluation of precision-oriented thresholding. We find that on many relations, GPT-3 can reproduce existing statements at over 95% precision, and there are significant gains over the smaller BERT-large model. At the same time, it should be noted that (Sun et al., 2023) observed that for large enough models, parameter scaling does not improve performance further, so it is well possible that these scores represent a ceiling w.r.t. model size.

Is this cost-effective? Previous works have estimated the cost of KB statement construction at 1 ct. (highly automated infobox scraping) to $2 (manual curation) (Paulheim, 2018). Based on our prompt size (avg. 174 tokens), the cost of one query is about 0.35 ct., with filtering increasing the cost per retained statement to about 0.7 ct. So LM prompting is monetarily competitive to previous infobox scraping works, though with much higher recall potential.

In absolute terms, prompting GPT-3 for all 48M incomplete subject-relation pairs reported in Table 2 would amount to an expense of $168,000, and yield approximately 27M novel statements.

Does “Don’t know” prompting help? In Table 4 (middle) we show the impact of using examples without an answer. The result is unsystematic, with notable gains in several relations, but some losses in others. Further research on calibrating model confidences seems important (Jiang et al., 2021; Singhania et al., 2023).

Table 3: Automated evaluation in the precision-thresholding setting: GPT-3 (text-davinci-003 with 175B parameters, text-curie-001 with 6.7B parameters) and BERT-large (340M parameters) on 1000 samples/relation from WD-Known.

Table 4: Effect of variations to the standard prompting setting.

Does textual context help? Table 4 (right) shows the results for prompting with context. Surprisingly, this consistenly made performance worse, with hardly any recall beyond 90% precision. This is contrary to earlier findings like (Petroni et al., 2020) (for BERT) or (Mallen et al., 2023) (for QA), who found that context helps, especially in the long tail. Our analysis indicates that, in the high-precision bracket, misleading contexts cause more damage (lead to high confidence in incorrect answers), than what helpful contexts do good (boost correct answers).

How many few-shot examples should one use? Few-shot learning for KBC works with remarkably few examples. While our default experiments, fol- lowing (Cohen et al., 2023), used 8 examples, we found actually no substantial difference to smaller example sizes as low as 4.

5        Conclusion

We provided the first analysis of the real KB completion potential of GPT. Our findings indicate that GPT-3 could add novel knowledge to Wikidata, at unprecedented scale and quality (27M statements at 90% precision). Compared with other approaches the estimated cost of $168,000 is surprisingly cheap, and well within the reach of industrial players. We also find that, in the high-precision bracket, GPT-3 distills web content to a degree that context augmentation does not easily help.

Open issues remain in particular around identifying high-confidence predictions within an LM’s generations (Jiang et al., 2021; Singhania et al., 2023; Xiong et al., 2023), and the choice of examples.

Limitations

Using LMs for automated knowledge generation comes with the standard risk of exacerbating de- mographic biases. For example, many of the best- performing relations are language-related, where the model presumably often estimates a person’s native language entirely from their name.

In terms of reproducibility, it should be noted that our results are tied to a closed-source commercial API. Although GPT-3/4/chatGPT are widely used in research and industry, and OpenAI has announced plans to keep stable model versions online, long-term reproducibility is not ensured, and the internal workings of GPT are publicly not known. Although statement generation is at the core of KB completion, for a complete KBC pipeline, we are still missing critical components. This concerns in particular entity disambiguation, which is essential for relations with more than a few hundred possible object values. Similarly, Wikidata and other KBs give critical importance to scrutable referencing of statements. This is not easily possible with LMs.

References

Dimitrios Alivanistos, Selene Báez Santamaría, Michael Cochez, Jan-Christoph Kalo, Emile van Krieken, and Thiviyan Thanapalasingam. 2023. Prompting as probing: Using language models for knowledge base construction.

Sören Auer, Christian Bizer, Georgi Kobilarov, Jens Lehmann, Richard Cyganiak, and Zachary G. Ives. 2007. DBpedia: A nucleus for a web of open data. ISWC.

Antoine Bordes, Nicolas Usunier, Alberto Garcia-Duran, Jason Weston, and Oksana Yakhnenko. 2013. Translating embeddings for modeling multi- relational data. neurIPS.

Nicola De Cao, Wilker Aziz, and Ivan Titov. 2021. Editing factual knowledge in language models. In EMNLP.

Andrew Carlson, Justin Betteridge, Bryan Kisiel, Burr Settles, Estevam R Hruschka, and Tom M Mitchell. 2010. Toward an architecture for never-ending lan- guage learning. In AAAI.

Roi Cohen, Mor Geva, Jonathan Berant, and Amir Globerson. 2023. Crawling the internal knowledge- base of language models. Findings of EACL.

Yanai Elazar, Nora Kassner, Shauli Ravfogel, Abhilasha Ravichander, Eduard Hovy, Hinrich Schütze, and Yoav Goldberg. 2021. Measuring and improving consistency in pretrained language models. TACL.

Benjamin Heinzerling and Kentaro Inui. 2021. Lan- guage models as knowledge bases: On entity repre- sentations, storage capacity, and paraphrased queries. In EACL.

Zhengbao Jiang, Jun Araki, Haibo Ding, and Graham Neubig. 2021. How can we know when language models know? on the calibration of language models for question answering. TACL.

Douglas B Lenat. 1995. Cyc: A large-scale investment in knowledge infrastructure. CACM.

Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2022. Pre- train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM CSUR.

Alex Mallen, Akari Asai, Victor Zhong, Rajarshi Das, Daniel Khashabi, and Hannaneh Hajishirzi. 2023. When not to trust language models: Investigating effectiveness of parametric and non-parametric mem- ories. In ACL.

Maximilian Nickel, Volker Tresp, and Hans-Peter Kriegel. 2011. A three-way model for collective learning on multi-relational data. In ICML.

Heiko Paulheim. 2018. How much is a triple? estimat- ing the cost of knowledge graph creation. In ISWC.

Fabio Petroni, Patrick Lewis, Aleksandra Piktus, Tim Rocktäschel, Yuxiang Wu, Alexander H Miller, and Sebastian Riedel. 2020. How context affects lan- guage models’ factual predictions. In AKBC.

Fabio Petroni, Tim Rocktäschel, Sebastian Riedel, Patrick Lewis, Anton Bakhtin, Yuxiang Wu, and Alexander Miller. 2019. Language models as knowl- edge bases? In EMNLP.

Nina Poerner, Ulli Waltinger, and Hinrich Schütze. 2020. E-BERT: Efficient-yet-effective entity embeddings for BERT. In Findings of EMNLP.

Simon Razniewski, Andrew Yates, Nora Kassner, and Gerhard Weikum. 2021. Language models as or for knowledge bases. DL4KG.

Adam Roberts, Colin Raffel, and Noam Shazeer. 2020. How much knowledge can you pack into the parame- ters of a language model? In EMNLP.

Tara Safavi and Danai Koutra. 2020. CoDEx: A com- prehensive knowledge graph completion benchmark. In EMNLP.

Taylor Shin, Yasaman Razeghi, Robert L. Logan IV, Eric Wallace, and Sameer Singh. 2020. AutoPrompt: Eliciting knowledge from language models with au- tomatically generated prompts. In EMNLP.

Sneha Singhania, Tuan-Phong Nguyen, and Simon Razniewski. 2022. LM-KBC: Knowledge base con- struction from pre-trained language models. CEUR.

Sneha Singhania, Simon Razniewski, and Gerhard Weikum. 2023. Extracting multi-valued relations from language models. Repl4NLP.

Fabian M Suchanek, Gjergji Kasneci, and Gerhard Weikum. 2007. Yago: a core of semantic knowledge. In WWW.

Kai Sun, Yifan Ethan Xu, Hanwen Zha, Yue Liu, and Xin Luna Dong. 2023. Head-to-tail: How knowl- edgeable are large language models (llm)? aka will llms replace knowledge graphs? arXiv.

Blerta Veseli, Sneha Singhania, Simon Razniewski, and Gerhard Weikum. 2023. Evaluating language models for knowledge base completion. In ESWC.

D. Vrandecˇic´ and M. Krötzsch. 2014. Wikidata: A free collaborative knowledge base. CACM.

Gerhard Weikum, Luna Dong, Simon Razniewski, and Fabian M. Suchanek. 2021. Machine Knowledge: Creation and Curation of Comprehensive Knowledge Bases. In FnT.

Miao Xiong, Zhiyuan Hu, Xinyang Lu, Yifei Li, Jie Fu, Junxian He, and Bryan Hooi. 2023. Can llms express their uncertainty? an empirical evaluation of confidence elicitation in llms. arXiv.

Yuqi Zhu, Xiaohan Wang, Jing Chen, Shuofei Qiao, Yixin Ou, Yunzhi Yao, Shumin Deng, Huajun Chen, and Ningyu Zhang. 2023. Llms for knowledge graph construction and reasoning: Recent capabilities and future opportunities. arXiv.

A        Predictions

In the following we show some predictions from our different experimental setups. Q marks the query, P the prediction by GPT-3 and A the ground truth object. Please note that to prompt GPT-3, relation-specific examples are used at the beginning of each query. The prompts w/ examples can be seen in Appendix B.

A.1         Standard

This section shows predictions from our standard setup with the number of few-shot examples being 8. The prompts B.1.

A.1.1        GPT-3

Q: As It Is in Heaven # original language

P: Swedish

A: Swedish

Q: John Gruber # employer

P: Daring Fireball

A: The Pittsburg Times, Pittsburgh Post, Pittsburgh Gazette, The Gazette Times, Pittsburgh Pi- rates

Q: Hillary Canyon # continent

P: Antarctica

A: North America

Q: The Incredible Hulk # developer

P: Edge of Reality

A: Edge of Reality, Amaze Entertainment

Q: Architecture and secondary structure of an entire HIV-1 RNA genome # language of work

P: English

A: English

A.1.2        GPT-4

Q: Who are the developers of The Incredible Hulk? Give me a list with no additional text.

P: Edge of Reality

A: Radical Entertainment, Edge of Reality, Amaze Entertainment, Pocket Studios, Universal Inter- active

Q:In which languages is Architecture and secondary structure of an entire HIV-1 RNA genome available? Give me a list with no additional text.

P: English

A: I´m sorry, but I couldn’t find specific information about the languages in which the mentioned scientific study or information is available.

A.1.3        ChatGPT

Q: Who are the developers of The Incredible Hulk? Give me a list with no additional text.

P: Edge of Reality

A: Edge of Reality, Amaze Entertainment, Beenox, Foundation 9 Entertainment, Griptonite Games, Fizz Factor, Nucore

Q:In which languages is In which languages is Architecture and secondary structure of an entire HIV-1 RNA genome available? Give me a list with no additional text. available? Give me a list with no additional text.

P: English

A: English

A: Kingdom of Italy, Italy

A.2        Textual context

In this section we show some predictions from our standard setup with textual contexts. The examples used for prompting in this setup can be seen in section B.2.

Q: Neko-Nin exHeart # developer C: Apr 21, 2017 … Developer. Whirlpool ; Publisher. Sekai Project ; Released. Apr 21, 2017; OS: Windows 7 or above ; Pro- cessor: 1.2 GHz Pentium 4

P: Whirlpool, Sekai Project

A: Whirlpool

Q: Navel*Plus # developer

C: Position: Qlik Sense Devel- oper. Location: Phoenix, AZ … SSIS and Datastage are a plus. Thanks & Regards,. Sravan Kumar | Navel Technologies Inc.

P: Navel Technologies Inc.

A: Navel

Q: Andrei Krasilnikov # native language

C: Languages · Russian. Native or bilingual proficiency · English. Full professional proficiency · French. Limited working proficiency · Spanish. Limited working

P: Spanish, French, English, Rus- sian

A: Russian

Q: Sergio Aquilante # citizen- ship

C: View Ariana Morais’ profile on LinkedIn, the world’s largest professional community. Ariana has 1 job listed on their profile. See the complete profile on

P: Brazil

Q: Luis Alfredo López # citizenship

C: Congratulations to Luis “Alfredo” Lopez, an associate from Vistar – Southern California, for being inducted into the 2020 IFDA Truck Driver Hall of Fame

P: Mexico

A: Honduras

A.3         Don’t Know

In this section we present some predictions from our setup using Don’t know prompting. The examples used for prompting in this setup can be seen in section B.3.

Q: Visvanath Kar # native lan- guage

P: Don’t know

A: Odia

Q: Eikeviken # manufacturer

P: Don’t know

A: Samsung Heavy Industries

Q: Volkwin Marg # work location

P: Hamburg, Aachen

A: Berlin, Brunswick, Hamburg, Aachen

Q: Elvive # manufacturer

P: L’Oréal

A: L’Oréal

A.4        Out-of-KB facts

In this section we present some predictions for out- of-KB facts evaluated in a manual evaluation. For prompting we used the examples from our standard setup B.1.

Q: Lalith Perera # citizenship

P: Sri Lanka

Q: The Splatters # developer

P: SpikySnail Games

Q: Perú Negro # location of formation

P: Lima

Q: Albrecht von Quadt # languages spoken, written or signed in

P: German

Q: Wreckless: The Yakuza Missions # developer

P: Bunkasha Games

B        Prompts

B.1         Standard employedBy

Q: Silvestre Paredes # employer

A: Universidad Politécnica de Cartagena

Q: Masashi Kamogawa # employer

A: Waseda University

Q: Ana Rosa Rama Ballesteros # employer

A: University of Jaén # Univer-

sity of Jaén

Q: Sara Akbar # employer A: Kuwait Oil Company Q: Pius V # employer

A: University of Bologna # University of Pavia

Q: Masao Kotani # employer

A: University of Tokyo # Osaka University

Q: Bernadeta Patro Golab # employer

A: Medical University of Warsaw

Q: Andy Hertzfeld # employer

A: Google developedBy

Q: Chronicles of Mystery: The Scorpio Ritual # developer

A:CI Games S.A.

Q: Call of Duty 3 # developer

A: Treyarch # Exakt Entertainment

Q: Samplitude # developer

A: Bellevue Investments

Q: Dangun Feveron # developer

A: CAVE

Q: Sega Classics Arcade Collection # developer

A: Sega

Q: Python # developer

A: Python Software Foundation # Guido van Rossum

Q: Allen Coral Atlas # developer

A: Vulcan Inc. # University of Queensland # Carnegie Institution for Science

Q: Avatar: The Last Airbender – Into the Inferno # developer

A: Nickelodeon

nativeLanguage

Q: Bill Byrge # native language

A: English

Q: Augustin Michel # native language

A: French

Q: Ingrian Finns # native language

A: Finnish # Russian

Q: Nanu Ram Kumawat # native language

A: Rajasthani # Hindi

Q: Jumber Dochviri # native language

A: Georgian

Q: Rosa Estaràs #native language

A: Spanish # Catalan

Q: Vladimir Fotievich Kozlov # native language

A: Russian

Q: Vladimir Nemkin # native language

A: Russian # Ukrainian

B.2       Textual context inContinent

Q: Reventador # continent

C: Daily explosions, ash plumes, lava flows, and incandescent block avalanches during February-July 2022. Volcán El Reventador is located 100 km E of the main… Reventador is an active stratovolcano which lies in the eastern Andes of Ecuador. It lies in a remote area of the national park of the same name, which is…

A: South America # Americas

Q: Fatimid Caliphate # continent

C: The Fatimid Caliphate was an Ismaili Shi’a caliphate extant from the tenth to the twelfth … Fatimid Caliphate is located in Continental Asia. Encompassing the vast Sahara in North Africa, alongside the Levant in the Middle East, the entirety of the Caliphate consists of Arid biome, save for Sicily and…

A: Asia # Africa # Europe

Q: Cerro Tenán # continent

C: See photos, floor plans and more details about 262 S Paseo Cerro in Green Valley, Arizona. Visit Rent. now for rental rates and other information about this… 200 N Continental Blvd … NEC Market St & Via Cerro … Investment Services · Landlord Representation · Tenant Representation · Industrial: Warehouse &…

A: Americas

Q: Mississippi River # continent C: Feb 10, 2022 … The Mississippi River is the second longest river in North America, flowing 2,350 miles from its source at Lake Itasca through the center of the… … or about one-eighth of the entire continent. The Mississippi River lies entirely within the United States. Rising in Lake Itasca in Minnesota, it flows…

A: Americas # North America

Q: St. Lawrence River # continent

C: The St. Lawrence River is a large river in the middle lat- itudes of North America. Its headwaters begin flowing from Lake Ontario in a roughly north- easterly… St. Lawrence River, hydrographic system of east- central North America. It starts at the outflow of Lake Ontario and leads into the Atlantic Ocean in the…

A: North America

Q: Cerro El Charabón # continent C: 65, Estancia El Charabón. 49. 66, Área costero-marina Cerro Verde e Islas de la Coronilla– Área General. 48. 67, Area protegida Laguna de Castillos – Tramo… Casa del Sol Boutique Hotel. A cozy stay awaits you in Machu Picchu. Altiplánico San Pedro de Atacama………… Welcome to El Charabon. El Charabon.

A: Americas

Q: Hinterer Seekopf # continent C: Following the breakup of Pangea during the Mesozoic era, the continents of………… of the best day hikes in Kalka- lpen National Park is the Ho- her Nock – Seekopf. Dec 5, 2016… Hinterer Steinbach. In- haltsverzeichnis aufklappen … Inhaltsverzeichnis einklappen … Charakteristik. Hinweise; Sub- jektive Bewertung…

A: Europe

Q: Šembera # continent

C: Rephrasing Heidegger: A Com- panion to Heidegger’s Being and Time [Sembera, … Being and Time (Suny Series in Contemporary Continental Philosophy). Feb 26, 2016 … Coming from Uganda, UNV PO Flavia Sembera was famil- iar with diversity…………………………….. shared across the continent while expe- riencing Zambia’s beautiful…

A: Europe work location

Q: Karel Lodewijk Sohl # work location

C: Karel Lodewijk SOHL geboren op 18 februari 1895 te Maastricht. Hij huwde Johanna Catharina Maria VAN BINSBERGEN 20 november 1924 te Roermond. Karel LANOO. Karen ANDERSON. Karina MROß. Karine LALIEUX … Lena SOHL. Leonidas MAKRIS. Leopold SPECHT … Lodewijk ASSCHER. Lora LYUBENOVA. Loren LANDAU …

A: Maastricht # Paris # Maastricht

Q: H.G. van Broekhuisen # work location

C: Nov 28, 2018 … Fourth, we are grateful for the work of the e-Vita platform helpdesks: Mireille … Van Spall HG, Rah- man T, Mytton O, Ramasundara- hettige C,… Apr 30, 2013 … Saskia F van Vugt, general practitioner 1,; Berna D L Broekhuizen, … In Flanders (Belgium) this work was supported by the Research…

A: Makassar

Q: Hermann Kretzschmer # work location

C: The author died in 1890, so this work is in the public domain in its country of origin and other countries and areas where the copyright term is the author’s… Stay up to date with Hermann Kretzschmer (German, 1811 – 1890) . Discover works for sale, auction results, market data, news and exhibitions on Mu- tualArt.

A: Berlin # Düsseldorf

Q: Ole Olsen Teslien # work loca- tion

C: og Bjom Pedersen Teslien 1831 – kjøpesum 3500 spd. Fra 1845 var Bjom Ptdersen alene eier. … Efter Ole Jensen overtok sønnen Anders Olsen g. i 1850. Jul 11, 2015 … Ole Olsen Bak eller Ormhaug som i 1803 kjøpte glrden ØvreViken 461, … Julius Sand. fra 1819 var Ole Olsen Teslien. someide Nedre n s li,…

A: Oslo

Q: John Sewell # work location C: John W Sewell was born in 1867 in Elbert County, Georgia, and moved with his parents to Florida when he was 19 years old. Sewell, working for Henry Flagler,. Implementation Consulting | Learn more about John Sewell, CMRP’s work experience, education, connections & more by visiting their profile on LinkedIn.

A: London # London

Q: Alan R. Battersby # work location

C: Sir Alan Rushton Battersby FRS (4 March 1925 – 10 February 2018) was an English organic chemist best known for his work to de- fine the chemical intermediates… Dec 9, 2018 … Battersby. Alan R. Battersby, University of Cam- bridge Cambridge, United Kingdom. “for their fundamental contri- butions to the elucidation of the…

A: Cambridge

Q: A. Kate Miller # work location

C: Title/Position. Advocacy Director. Department. Advocacy Department. Pronouns. she … Kate works in a number of issue areas and is always seeking com- mon ground… … Miller’s work experience, education, connec- tions & more by visiting their profile on LinkedIn………………………………… Work location: Chicago, Illinois, United States. Work…

A: Indianapolis

Q: Andrzej Grzesik # work location

C: View Andrzej Grzesik’s profile on LinkedIn, the world’s largest professional community. Andrzej has 15 jobs listed on their profile. Nov 14, 2016 … Congratulations to the newest Java Champion Andrzej Grzesik! … in Poland (sfi.org.pl) and in his work as a Sun Campus Ambassador.

A: Warsaw

B.3       Don’t Know producedBy

Q: Strati # manufacturer

A: Local Motors # Oak Ridge National Laboratory

Q: Philips VG-8235 # manufacturer

A: Don’t know v SS America (1939) # manufacturer

A: Newport News Shipbuilding

Q: HMCS Brandon # manufacturer

A: Don’t know

Q: Cluster Platform 3000 SL160z, Xeon L55xx 2.26 GHz, GigE # manufacturer

A: Hewlett Packard Enterprise

Q: POWER CHALLENGE # manufacturer

A: Don’t know

Q: Suzuki Katana # manufacturer

A: Suzuki

Q: German submarine U-1223 # manufacturer

A:Don’t know

spokenLanguage

Q: Allen G. Thurman # languages spoken, written or signed

A: English

Q: Rifat Hairy # languages spo- ken, written or signed

A: Don’t know

Q: Izabela Filipiak # languages spoken, written or signed

A: American English # Polish

Q: Vasudev Gopal Paranjpe # lan- guages spoken, written or signed

A: Don’t know

Q: Jonathan M. Katz # languages spoken, written or signed

A: English

Q: Ingeborg Heintze # languages spoken, written or signed

A: Don’t know

Q: Alicia Coduras Martínez # lan- guages spoken, written or signed

A: Catalan # Spanish

Q: Adam Budak # languages spoken, written or signed

A: Don’t know