diff --git a/freebase/ComplexQuestions.md b/freebase/ComplexQuestions.md
index 9837f551..8b265d08 100644
--- a/freebase/ComplexQuestions.md
+++ b/freebase/ComplexQuestions.md
@@ -5,6 +5,7 @@
| Model / System | Year | Average F1 | Reported by |
|:-------------------:|:---------:|:----------:|:----------------------------------------------------------------------------:|
+| TFS-KBQA | 2024 | 44.0 | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf)|
| Lan and Jiang [1] | 2020 | 43.3 | [Yonghui Jia and Wenliang Chen](https://arxiv.org/pdf/2204.12808.pdf) |
| Reranking | 2022 | 42.9 | [Yonghui Jia and Wenliang Chen](https://arxiv.org/pdf/2204.12808.pdf) |
| Luo et al. [2] | 2020 | 42.8 | [Yonghui Jia and Wenliang Chen](https://arxiv.org/pdf/2204.12808.pdf) |
diff --git a/freebase/ComplexWebQuestions.md b/freebase/ComplexWebQuestions.md
index 5752d67f..d53144e8 100644
--- a/freebase/ComplexWebQuestions.md
+++ b/freebase/ComplexWebQuestions.md
@@ -13,19 +13,30 @@
| GPT-3 | 2023 | - | - | 51.77 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
| FLAN-T5 | 2023 | - | - | 46.69 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
| BART-large | 2022 | 68.2 | - | - | EN | [Hu et al.](https://aclanthology.org/2022.coling-1.145.pdf) |
+| FiDeLiS + gpt-4-turbo | 2024 | 64.32| 71.47 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| TFS-KBQA | 2024 | 63.6 | - | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
| DECAF (BM25 + FiD-3B) | 2022 | - | 70.4 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| CBR-KBQA | 2022 | - | 70.4 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| DECAF (BM25 + FiD-large) | 2022 | - | 68.1 ± 0.5 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| DECAF (DPR + FiD-large) | 2022 | - | 67.0 ± 0.4 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| FiDeLiS + gpt-3.5-turbo | 2024 | 61.78| 63.12 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| ToG + gpt-4-turbo | 2024 | 60.2 | 68.51 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| HGNet | 2022 | - | 58.1 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| ProgramTransfer | 2022 | 58.7 | 58.1 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
+| ToG + gpt-3.5-turbo | 2024 | 56.96| 57.59 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| RoG | 2024 | 56.17| 61.39 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| ProgramTransfer-o | 2022 | 55.8 | 54.7 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
| ProgramTransfer-pa | 2022 | 54.5 | 54.3 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
+| KD-CoT | 2024 | - | 53.92 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| NSM+h | 2022 | - | 53.9 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| FC-KBQA | 2024 | 54.1 | - | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
| REAREV | 2022 | - | 52.9 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
| QNRKGQA+h | 2022 | - | 51.5 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
| SR+NSM | 2022 | - | 50.2 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| QNRKGQA | 2022 | - | 50.5 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
+| RCAANet | 2024 | - | 50.5 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
+| RE-KBQA | 2024 | - | 50.3 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
+| DGEN | 2024 | - | 49.5 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
| NSM-distill | 2022 | - | 48.8 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
| Rigel | 2022 | - | 48.7 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
| TransferNet | 2022 | - | 48.6 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
@@ -36,6 +47,7 @@
| TERP | 2022 | - | 49.2 | - | EN | [Qiao et al.](https://aclanthology.org/2022.coling-1.156.pdf) |
| TeacherNet | 2022 | 44.0 | 48.8 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
| NSM | 2022 | 44.0 | 48.8 | - | EN | [Du et al.](https://arxiv.org/pdf/2209.03005.pdf) |
+| RnG-KBQA | 2024 | 42.3 | - | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
| NSM+h | 2022 | - | 48.8 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
| TransferNet | 2022 | - | 48.6 | - | EN | [Du et al.](https://arxiv.org/pdf/2209.03005.pdf) |
| NSM+p | 2022 | - | 48.3 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
diff --git a/freebase/WebQuestionsSP.md b/freebase/WebQuestionsSP.md
index 0b99b378..7e781a96 100644
--- a/freebase/WebQuestionsSP.md
+++ b/freebase/WebQuestionsSP.md
@@ -6,11 +6,17 @@
| Model / System | Year | F1 | Hits@1 | Accuracy | Language | Reported by |
| :---------------------------------: | :--: | :--------: | :--------: | :------: | :------: | :-----------------------------------------------------------------------------------: |
| chatGPT | 2023 | - | - | 83.70 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
+| TFS-KBQA | 2024 | 79.9 | 79.8 | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
+| Pangu | 2024 | 79.6 | - | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
| TIARA | 2022 | 78.9 | 75.2 | - | EN | [Shu et. al.](https://aclanthology.org/2022.emnlp-main.555.pdf) |
| DECAF (DPR + FiD-3B) | 2022 | 78.8 | 82.1 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| FiDeLiS + gpt-4-turbo | 2024 | 78.32 | 84.39 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| GPT-3.5v3 | 2023 | - | - | 79.60 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
| DECAF (DPR + FiD-large) | 2022 | 77.1 ± 0.2 | 80.7 ± 0.2 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| FC-KBQA | 2024 | 76.9 | - | - | EN | [Shouhui Wang and Biao Qin](https://aclanthology.org/2024.lrec-main.1074.pdf) |
+| FiDeLiS + gpt-3.5-turbo | 2024 | 76.78 | 79.32 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| UniK-QA | 2022 | - | 79.1 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| Sun et al. | 2024 | - | 77.8 | - | EN | [Sun et al.](https://aclanthology.org/2024.lrec-main.496.pdf) |
| TERP | 2022 | - | 76.8 | - | EN | [Qiao et al.](https://aclanthology.org/2022.coling-1.156.pdf) |
| SQALER+GNN | 2022 | - | 76.1 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
| EmQL | 2020 | - | 75.5 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
@@ -18,6 +24,7 @@
| GPT-3.5v2 | 2023 | - | - | 72.34 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
| Program Transfer | 2022 | 76.5 | 74.6 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| RnG-KBQA (T5-large) | 2022 | 76.2 ± 0.2 | 80.7 ± 0.2 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| ToG + gpt-4-turbo | 2024 | 75.97 | 81.84 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| RnG-KBQA | 2022 | 75.6 | - | 71.1 | EN | [Hu et al.](https://aclanthology.org/2022.coling-1.145.pdf) |
| ArcaneQA | 2022 | 75.3 | - | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| QNRKGQA+h | 2022 | - | 75.7 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
@@ -26,6 +33,7 @@
| QNRKGQA | 2022 | - | 74.9 | - | EN | [Ma et al.](https://link.springer.com/chapter/10.1007/978-3-031-10983-6_11) |
| ReTrack | 2022 | 74.7 | - | - | EN | [Hu et al.](https://aclanthology.org/2022.coling-1.145.pdf) |
| ReTrack | 2021 | 74.6 | 74.7 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| DGEN | 2024 | - | 74.5 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
| BART-large | 2022 | 74.6 | - | - | EN | [Hu et al.](https://aclanthology.org/2022.coling-1.145.pdf) |
| Subgraph Retrieval | 2022 | 74.5 | 83.2 | - | EN | [Shu et. al.](https://aclanthology.org/2022.emnlp-main.555.pdf) |
| QGG | 2022 | 74.0 | - | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
@@ -52,6 +60,8 @@
| NSM | 2021 | - | 74.30 | - | EN | [He et al.](https://arxiv.org/pdf/2101.03737.pdf) |
| Rigel | 2022 | - | 73.3 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
| SGM | 2022 | 72.36 | - | - | EN | [Ma L et al.](https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9747229) |
+| ToG + gpt-3.5-turbo | 2024 | 72.32 | 75.13 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| RCAANet | 2024 | 72.3 | 76.4 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
| CBR-SUBG | 2022 | 72.1 | - | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| NPI | 2022 | - | 72.6 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
| TextRay | 2022 | - | 72.2 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
@@ -66,8 +76,10 @@
| HGNet | 2021 | 70.3 | 70.6 | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| GrailQA Ranking | 2021 | 70.0 | - | - | EN | [Shu et. al.](https://aclanthology.org/2022.emnlp-main.555.pdf) |
| SQALER | 2022 | - | 70.6 | - | EN | [Costas Mavromatis and George Karypis](https://arxiv.org/pdf/2210.13650.pdf) |
+| RoG | 2024 | 69.81 | 83.15 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| STAGG | 2015 | 69.00 | - | - | EN | [Ma L et al.](https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9747229) |
| UHop | 2019 | 68.5 | - | - | EN | [Ma L et al.](https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9747229) |
+| RE-KBQA | 2024 | 68.5 | 74.6 | - | EN | [Lv et al.](https://www.researchsquare.com/article/rs-4292912/v1) |
| KBIGER | 2022 | 68.4 | 75.3 | - | EN | [Du et al.](https://arxiv.org/pdf/2209.03005.pdf) |
| NSM | 2022 | - | 69.0 | - | EN | [Cao et al.](https://aclanthology.org/2022.acl-long.559.pdf) |
| GraftNet-EF+LF | 2018 | - | 68.7 | - | EN | [Sun et al.](https://aclanthology.org/D18-1455.pdf) |
@@ -96,6 +108,7 @@
| FLAN-T5 | 2023 | - | - | 59.87 | EN | [Tan et al.](https://arxiv.org/pdf/2303.07992.pdf) |
| KGT5 | 2022 | 56.1 | - | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| FILM | 2022 | 54.7 | - | - | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
+| KD-CoT | 2024 | 50.2 | 73.7 | - | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
| ReifKB | 2020 | - | 52.7 | - | EN | [Cohen et al.](https://arxiv.org/pdf/2002.06115.pdf) |
| KV-Mem | 2022 | 38.6 | 46.7 | - | EN | [Du et al.](https://arxiv.org/pdf/2209.03005.pdf) |
| KGQA-RR(GPT2) | 2023 | - | - | 18.11 | EN | [Hu et al.](https://arxiv.org/pdf/2303.10368.pdf) |
diff --git a/freebase/freebaseQA.md b/freebase/freebaseQA.md
index 697917b6..cebce305 100644
--- a/freebase/freebaseQA.md
+++ b/freebase/freebaseQA.md
@@ -5,6 +5,7 @@
| Model / System | Year | Exact Match | Accuracy | Hits@1 | Language | Reported by |
| :----------------------: | :--: | :---------: | :------: | :-------: | :------: | :----------------------------------------------------: |
+| Sun et al. | 2024 | - | - | 84.3 | EN | [Sun et al.](https://aclanthology.org/2024.lrec-main.496.pdf) |
| DECAF (DPR + FiD-large) | 2022 | - | - | 79.0±0.6 | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| DECAF (BM25 + FiD-large) | 2022 | - | - | 78.8± 0.5 | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
| FILM | 2022 | - | - | 63.3 | EN | [Yu et al.](https://arxiv.org/pdf/2210.00063.pdf) |
diff --git a/frontend/static/entries.txt b/frontend/static/entries.txt
index 48f6ca4d..581e9bb9 100644
--- a/frontend/static/entries.txt
+++ b/frontend/static/entries.txt
@@ -55,6 +55,7 @@
/datasets/other/
/datasets/other/COVID-QA
/datasets/other/CrunchQA
+/datasets/other/CR-LT-KGQA
/datasets/other/DBLP-QuAD
/datasets/other/Event-QA
/datasets/other/MetaQA - 1 Hop
diff --git a/other/CR-LT-KGQA.md b/other/CR-LT-KGQA.md
new file mode 100644
index 00000000..7b0346df
--- /dev/null
+++ b/other/CR-LT-KGQA.md
@@ -0,0 +1,18 @@
+---
+ name: CR-LT-KGQA
+ datasetUrl: https://github.com/D3Mlab/cr-lt-kgqa
+---
+
+| Model / System | Year | Accuracy | Language | Reported by |
+|:-----------------------------------:|:----:|:--------:|:--------:|:----------------------------------------------------------------------------------:|
+| gpt-3.5-turbo (Zero-shot) | 2024 | 32.74 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-3.5-turbo (Few-shot) | 2024 | 36.61 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-3.5-turbo (Chain of Thought) | 2024 | 37.42 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-4-turbo (Zero-shot) | 2024 | 37.74 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-4-turbo (Few-shot) | 2024 | 43.61 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-4-turbo (Chain of Thought) | 2024 | 45.42 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| Reasoning on Graphs (RoG) | 2024 | 60.32 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-3.5-turbo + ToG | 2024 | 62.48 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-3.5-turbo + FiDeLiS | 2024 | 67.34 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-4-turbo + ToG | 2024 | 67.24 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
+| gpt-4-turbo + FiDeLiS | 2024 | 72.12 | EN | [Sui et al.](https://arxiv.org/pdf/2405.13873) |
\ No newline at end of file
diff --git a/other/^CR-LT-KGQA.md b/other/^CR-LT-KGQA.md
new file mode 100644
index 00000000..e4b8c501
--- /dev/null
+++ b/other/^CR-LT-KGQA.md
@@ -0,0 +1,9 @@
+# CR-LT-KGQA
+
+**CR-LT-KGQA**[[1]](#myfootnote1) is a Knowledge Graph Query Answering dataset with Natural Language queries targeted on long-tail Wikidata entities, answering which requires commonsense reasoning and is submitted to SIGIR'24. The dataset contains two subsets targetting two different tasks: (i) Question Answering subset containing 200 questions based on StrategyQA dataset and (ii) Claim Verification subset containing 150 claims based on Creak dataset.
+
+The format of the dataset is in JSON, where each entry contains a query (a question or a claim), the answer, anchor KG entities mentioned in the query and their respective Wikidata QID, an inference rule, relevant KG triples, reasoning steps and the relevant KG triples to each step, and finally the set of reasoning skills and strategies required to answer the query.
+
+
+## References
+[1] Guo, Willis and Toroghi, Armin and Sanner, Scott. [CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring Commonsense Reasoning and Long-Tail Knowledge](https://arxiv.org/pdf/2403.01395) arXiv preprint arXiv:2403.01395. 2024.
diff --git a/systems.md b/systems.md
index b5dbc723..73f5dacc 100644
--- a/systems.md
+++ b/systems.md
@@ -141,4 +141,23 @@
| KGQAcl/rr | Hu et al. | [Link](https://arxiv.org/pdf/2303.10368.pdf) | yes | [Link](https://github.com/HuuuNan/PLMs-in-Practical-KBQA) | [Link](https://arxiv.org/pdf/2303.10368.pdf) | KGQA-CL and KGQA-RR are tow frameworks proposed to evaluate the PLM's performance in comparison to their efficiency. Both architectures are composed of mention detection, entity disambiguiation, relation detection and answer query building. The difference lies on the relation detection module. KGQA-CL aims to map question intent to KG relations. While KGQA-RR ranks the related relations to retrieve the answer entity. Both frameworks are tested on common PLM, distilled PLMs and knowledge-enhanced PLMs and achieve high performance on three benchmarks. | Hu et al. |
| W. Han et al. | Han et al. | [Link](https://link.springer.com/chapter/10.1007/978-3-031-30672-3_39) | no | - | [Link](https://link.springer.com/chapter/10.1007/978-3-031-30672-3_39) | This model is based on machine reading comprehension. To transform a subgraph of the KG centered on the topic entity into text, the subgraph is sketched through a carefully designed schema tree, which facilitates the retrieval of multiple semantically-equivalent answer entities. Then, the promising paragraphs containing answers are picked by a contrastive learning module. Finally, the answer entities are delivered based on the answer span that is detected by the machine reading comprehension module. | Han et al. |
| GAIN | Shu et al. | [Link](https://arxiv.org/pdf/2309.08345.pdf) | no | - | [Link](https://arxiv.org/pdf/2309.08345.pdf) | GAIN is not a KGQA system, but a data augmentation method named Graph seArch and questIon generatioN (GAIN). GAIN applies to KBQA corresponding to logical forms or triples, and scales data volume and distribution through four steps: 1) Graph search: Sampling logical forms or triples from arbitrary domains in the KB, without being restricted to any particular KBQA dataset. 2) Training question generator on existing KBQA datasets, i.e., learning to convert logical forms or triples into natural language questions. 3) Verbalization: Using the question generator from step 2 to verbalize sampled logical forms or triples from step 1, thus creating synthetic questions. 4) Training data expansion: Before fine-tuning any neural models on KBQA datasets, GAIN-synthetic data can be used to train these models or to expand the corpus of in-context samples for LLMs. That is, as a data augmentation method, GAIN is not a KBQA model, but it is used to augment a base KBQA model. | Shu et al. |
-| JarvisQALcs | Jaradeh et al. | [Link](https://arxiv.org/pdf/2006.01527) | no | | | same as reporting paper | JarvisQA a BERT based system to answer questions on tabular views of scholarly knowledge graphs. | Jaradeh et al. |
\ No newline at end of file
+| JarvisQALcs | Jaradeh et al. | [Link](https://arxiv.org/pdf/2006.01527) | no | | | same as reporting paper | JarvisQA a BERT based system to answer questions on tabular views of scholarly knowledge graphs. | Jaradeh et al. |
+| DGEN | Lv et al. | [Link](https://www.researchsquare.com/article/rs-4292912/v1) | no | - | [Link](https://www.sciencedirect.com/science/article/pii/S0925231222012905?ref=pdf_download&fr=RR-2&rr=892a7e715f3c62f5) | The Dynamic Graph Expansion Network (DGEN) is a network that forms a joint subgraph between the question and entities for reasoning, where the whole question is regarded as a node of the subgraph. At each step we believe that there is reasoning related information in the context of the question which will affect the auxiliary entity and intermediate path selection. To this end, we design a dynamic subgraph module and a joint reasoning module. The former is intended to dynamically adjust the subgraph in light of the question context. Built on top of that, the latter is intended to combine the interaction between the dynamic subgraph and intermediate supervision signal for joint reasoning. Through this idea of dynamic subgraph joint reasoning, we are able to distill the information that is the most relevant to reasoning the answer on both sides of the question context and knowledge base (KB). | Wu et al. |
+| RE-KBQA | Lv et al. | [Link](https://www.researchsquare.com/article/rs-4292912/v1) | yes | [Link](https://github.com/yongcaoplus/RE-KBQA) | [Link](https://arxiv.org/pdf/2305.02118) | RE-KBQA, utilizes relations in the knowledge base to enhance entity representation and introduce additional supervision. We explore guidance from relations in three aspects, including (1) distinguishing similar entities by employing a variational graph auto-encoder to learn relation importance; (2) exploring extra supervision by predicting relation distributions as soft labels with a multi-task scheme; (3) designing a relation-guided re-ranking algorithm for post-processing. | Cao et al. |
+| ReifKB | Lv et al. | [Link](https://www.researchsquare.com/article/rs-4292912/v1) | yes | [Link](https://github.com/google-research/language) | [Link](https://arxiv.org/pdf/2002.06115) | We describe a novel way of representing a symbolic knowledge base (KB) called a sparse-matrix reified KB. This representation enables neural KB inference modules that are fully differentiable, faithful to the original semantics of the KB, expressive enough to model multi-hop inferences, and scalable enough to use with realistically large KBs. The sparse-matrix reified KB can be distributed across multiple GPUs, can scale to tens of millions of entities and facts, and is orders of magnitude faster than naive sparse-matrix implementations. | Cohen et al. |
+| RCAANet | Lv et al. | [Link](https://www.researchsquare.com/article/rs-4292912/v1) | no | - | [Link](https://www.researchsquare.com/article/rs-4292912/v1) | This paper proposes a new multi-hop model, called Relation Crossfusion Attention Assistance Networks(RCAANet), which has the following advantages: 1) Accuracy. It can exploit the interactive information of question representation and relation embedding to help calculate the accurate relation distribution of each hop. 2) Effectiveness. RCAANet outperforms previous models, improving multi-hop question answering accuracy on two widely used complex question answering datasets. 3) Transparency. Similar to TransferNet, it is completely attention-based, so its intermediate steps are easily visualized and understood by humans. | Lv et al. |
+| Sun et al. | Sun et al. | [Link](https://aclanthology.org/2024.lrec-main.496.pdf) | no | - | [Link](https://aclanthology.org/2024.lrec-main.496.pdf) | Our pipeline (i) retrieves over documents of triples grouped by entity, (ii) re-ranks triples from these documents with context: triples in the 1-hop neighborhood of the documents’ subject entity, and (iii) generates an answer from highly relevant re-ranked triples. | Sun et al. |
+| Reasoning on Graphs (RoG) | Sui et al. | [Link](https://arxiv.org/pdf/2405.13873) | no | - | [Link](https://arxiv.org/pdf/2310.01061) | This paper proposes a novel method called reasoning on graphs (RoG) that synergizes LLMs with KGs to enable faithful and interpretable reasoning. Specifically, it presents a planning-retrieval-reasoning framework, where RoG first generates relation paths grounded by KGs as faithful plans. These plans are then used to retrieve valid reasoning paths from the KGs for LLMs to conduct faithful reasoning. Furthermore, RoG not only distills knowledge from KGs to improve the reasoning ability of LLMs through training but also allows seamless integration with any arbitrary LLMs during inference. | Luo et al. |
+| Think-on-Graph (ToG) | Sui et al. | [Link](https://arxiv.org/pdf/2405.13873) | yes | [Link](https://github.com/IDEA-FinAI/ToG) | [Link](https://arxiv.org/pdf/2307.07697) | This paper proposes a new LLM-KG integrating paradigm “LLM ⊗ KG ” which treats the LLM as an agent to interactively explore related entities and relations on KGs and perform reasoning based on the retrieved knowledge. The authors further implement this paradigm by introducing a new approach called Think-on-Graph (ToG), in which the LLM agent iteratively executes beam search on KG, discovers the most promising reasoning paths, and returns the most likely reasoning results. | Sun et al. |
+| KD-CoT | Sui et al. | [Link](https://arxiv.org/pdf/2405.13873) | yes | [Link](https://github.com/AdelWang/KD-CoT/tree/main) | [Link](https://arxiv.org/pdf/2308.13259) | This paper proposes a framework called Knowledge-Driven Chain-of-Thought (KD-CoT) to verify and modify reasoning traces in CoT via interaction with external knowledge, and thus overcome the hallucinations and error propagation. Concretely, the authors formulate the CoT rationale process of LLMs into a structured multi-round QA format. In each round, LLMs interact with a QA system that retrieves external knowledge and produce faithful reasoning traces based on retrieved precise answers. The structured CoT reasoning of LLMs is facilitated by our developed KBQA CoT collection, which serves as in-context learning demonstrations and can also be utilized as feedback augmentation to train a robust retriever. | Wang et al. |
+| FiDeLiS | Sui et al. | [Link](https://arxiv.org/pdf/2405.13873) | no | - | [Link](https://arxiv.org/pdf/2405.13873) | This paper proposes a retrieval-exploration interactive method, FiDeLiS to handle intermediate steps of reasoning grounded by KGs. Specifically, the authors propose Path-RAG module for retrieving useful intermediate knowledge from KG for LLM reasoning. FiDeLiS incorporates the logic and common-sense reasoning of LLMs and the topological connectivity of KGs into the knowledge retrieval process, which provides more accurate retrieving performance. Furthermore, FiDeLiS leverages deductive reasoning capabilities of LLMs as a better criterion to automatically guide the reasoningprocess in a stepwise and generalizable manner. Deductive verification serves as a precise indicator for when to cease further reasoning, thus avoiding misleading the chains of reasoning and unnecessary computation. | Sui et al. |
+| Subgraph Retrieval (SR) | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/RUCKBReasoning/SubgraphRetrievalKBQA) | [Link](https://aclanthology.org/2022.acl-long.396v2.pdf) | This paper introduces a trainable subgraph retriever (SR) for multi-hop Knowledge Base Question Answering (KBQA) that is decoupled from the reasoning process, allowing for better scalability and adaptability across various reasoning models. The SR model achieves significant improvements over existing methods by allowing a more targeted retrieval of subgraphs from large knowledge bases, reducing the reasoning space and bias. It employs dual-encoder architecture for efficient subgraph expansion, supports weakly supervised pre-training, and allows end-to-end fine-tuning with feedback from the reasoning model, enhancing overall QA performance. Extensive tests on benchmarks like WebQSP and CWQ show that SR not only improves retrieval effectiveness but also sets new performance standards for embedding-based KBQA systems. | Zhang et al. |
+| ReTraCk | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | no | - | [Link](https://aclanthology.org/2021.acl-demo.39/) | This paper presents Retriever-Transducer-Checker (ReTraCk), a neural semantic parsing framework for large scale knowledge base question answering (KBQA). ReTraCk is designed as a modular framework to maintain high flexibility. It includes a retriever to retrieve relevant KB items efficiently, a transducer to generate logical form with syntax correctness guarantees and a checker to improve the transduction procedure. | Chen et al. |
+| RnG-KBQA | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/salesforce/rng-kbqa) | [Link](https://aclanthology.org/2022.acl-long.417.pdf) | This paper presents RnG-KBQA, a Rank-and-Generate approach for KBQA, which remedies the coverage issue with a generation model while preserving a strong generalization capability. The approach first uses a contrastive ranker to rank a set of candidate logical forms obtained by searching over the knowledge graph. It then introduces a tailored generation model conditioned on the question and the top-ranked candidates to compose the final logical form. | Ye et al. |
+| ArcaneQA | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/dki-lab/ArcaneQA) | [Link](https://arxiv.org/pdf/2204.08109) | This paper presents ArcaneQA, a novel generation-based model that addresses both the large search space and the schema linking challenges in a unified framework with two mutually boosting ingredients: dynamic program induction for tackling the large search space and dynamic contextualized encoding for schema linking. Experimental results on multiple popular KBQA datasets demonstrate the highly competitive performance of ArcaneQA in both effectiveness and efficiency. | Gu and Su |
+| Program Transfer | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/THU-KEG/ProgramTransfer) | [Link](https://aclanthology.org/2022.acl-long.559/) | This paper proposes the approach of program transfer, which aims to leverage the valuable program annotations on the rich-resourced KBs as external supervision signals to aid program induction for the low-resourced KBs that lack program annotations. The authors design a novel two-stage parsing framework with an efficient ontology-guided pruning strategy. First, a sketch parser translates the question into a high-level program sketch, which is the composition of functions. Second, given the question and sketch, an argument parser searches the detailed arguments from the KB for functions. During the searching, we incorporate the KB ontology to prune the search space. | Cao et al. |
+| Pangu | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/dki-lab/Pangu) | [Link](https://aclanthology.org/2023.acl-long.270.pdf) | This paper proposes Pangu, a generic framework for grounded language understanding that capitalizes on the discriminative ability of LMs instead of their generative ability. Pangu consists of a symbolic agent and a neural LM working in a concerted fashion: The agent explores the environment to incrementally construct valid plans, and the LM evaluates the plausibility of the candidate plans to guide the search process. | Gu et al. |
+| FC-KBQA | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | no | - | [Link](https://aclanthology.org/2023.acl-long.57.pdf) | This paper proposes a Fine-to-Coarse Composition framework for KBQA (FC-KBQA) to both ensure the generalization ability and executability of the logical expression. The main idea of FC-KBQA is to extract relevant finegrained knowledge components from KB and reformulate them into middle-grained knowledge pairs for generating the final logical expressions. | Zhang et al. |
+| TFS-KBQA | Shouhui Wang and Biao Qin | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | yes | [Link](https://github.com/shouh/TFS-KBQA) | [Link](https://aclanthology.org/2024.lrec-main.1074.pdf) | This study adopts the LLMs, such as Large Language Model Meta AI (LLaMA), as a channel to connect natural language questions with structured knowledge representations and proposes a Three-step Fine-tune Strategy based on large language model to implement the KBQA system (TFS-KBQA). This method achieves direct conversion from natural language questions to structured knowledge representations, thereby overcoming the limitations of existing KBQA methods, such as addressing large search and reasoning spaces and ranking massive candidates. | Shouhui Wang and Biao Qin |
+| TSQA | Gao et al. | [Link](https://arxiv.org/pdf/2402.16568) | no | - | [Link](https://arxiv.org/pdf/2203.00255) | In this paper, we propose a timesensitive question answering (TSQA) framework to tackle these problems. TSQA features a timestamp estimation module to infer the unwritten timestamp from the question. We also employ a time-sensitive KG encoder to inject ordering information into the temporal KG embeddings that TSQA is based on. | Shang et al. |
+| GenTKGQA | Gao et al. | [Link](https://arxiv.org/pdf/2402.16568) | no | - | [Link](https://arxiv.org/pdf/2402.16568) | This paper first proposes a novel generative temporal knowledge graph question answering framework, GenTKGQA, which guides LLMs to answer temporal questions through two phases: Subgraph Retrieval and Answer Generation. First, we exploit LLM’s intrinsic knowledge to mine temporal constraints and structural links in the questions without extra training, thus narrowing down the subgraph search space in both temporal and structural dimensions. Next, we design virtual knowledge indicators to fuse the graph neural network signals of the subgraph and the text representations of the LLM in a non-shallow way, which helps the open-source LLM deeply understand the temporal order and structural dependencies among the retrieved facts through instruction tuning. | Gao et al. |
diff --git a/wikidata/CronQuestions.md b/wikidata/CronQuestions.md
index b6142404..8c58daad 100644
--- a/wikidata/CronQuestions.md
+++ b/wikidata/CronQuestions.md
@@ -5,9 +5,12 @@
| Model / System | Year | Hits@1 | Hits@10 | Reported by |
| :------------: | :--: | :----: | :-----: | :---------------------------------------------------------------------------: |
+| GenTKGQA | 2024 | 0.978 | 0.981 | [Gao et. al.](https://arxiv.org/pdf/2402.16568) |
| TempoQR-Hard | 2021 | 0.918 | 0.978 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
+| TSQA | 2024 | 0.831 | 0.980 | [Gao et. al.](https://arxiv.org/pdf/2402.16568) |
| TempoQR-Soft | 2021 | 0.799 | 0.957 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
| TMA | 2021 | 0.784 | 0.943 | [Liu et. al.](https://arxiv.org/pdf/2302.12529.pdf) |
+|ChatGPT (with TKG)| 2024 | 0.754 | 0.852 | [Gao et. al.](https://arxiv.org/pdf/2402.16568) |
| EntityQR | 2021 | 0.745 | 0.944 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
| CronKGQA | 2021 | 0.647 | 0.884 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
| EaE | 2021 | 0.288 | 0.678 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
@@ -15,6 +18,7 @@
| T-EaE-add | 2021 | 0.278 | 0.663 | [Saxena et. al.](https://arxiv.org/pdf/2106.01515.pdf) |
| BERT | 2021 | 0.243 | 0.620 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
| RoBERTa | 2021 | 0.225 | 0.585 | [Mavromatis et. al.](https://arxiv.org/pdf/2112.05785.pdf) |
+|ChatGPT (without TKG)| 2024 | 0.151 | 0.308 | [Gao et. al.](https://arxiv.org/pdf/2402.16568) |
| BERT | 2021 | 0.071 | 0.213 | [Saxena et. al.](https://arxiv.org/pdf/2106.01515.pdf) |
| RoBERTa | 2021 | 0.07 | 0.202 | [Saxena et. al.](https://arxiv.org/pdf/2106.01515.pdf) |
| KnowBERT | 2021 | 0.07 | 0.201 | [Saxena et. al.](https://arxiv.org/pdf/2106.01515.pdf) |