alexander2618 commited on
Commit
4ee29d7
·
1 Parent(s): afa90b9

update readme

Browse files
Files changed (1) hide show
  1. README.md +83 -3
README.md CHANGED
@@ -1,3 +1,83 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Agent4S-BioKG
2
+ A Knowledge Graph Checking Benchmark of AI Agent for Biomedical Science.
3
+ <p align="left">
4
+ <a href="https://github.com/westlake-autolab/Agent4S-BioKG/blob/main/LICENSE" alt="license">
5
+ <img src="https://img.shields.io/badge/license-MIT-blue" /></a>
6
+
7
+ <a href="https://github.com/westlake-autolab/Agent4S-BioKG" alt="license">
8
+ Github </a>
9
+ </p>
10
+
11
+ ## Introduction
12
+ Pursuing artificial intelligence for biomedical science, a.k.a. AI Scientist, draws increasing attention, where one common approach is to build a copilot agent driven by Large Language Models~(LLMs).
13
+ However, to evaluate such systems, people either rely on direct Question-Answering~(QA) to the LLM itself, or in a biomedical experimental manner. How to precisely benchmark biomedical agents from an AI Scientist perspective remains largely unexplored. To this end, we draw inspiration from one most important abilities of scientists, understanding the literature, and introduce `BioKGBench`.
14
+ In contrast to traditional evaluation benchmark that only focuses on factual QA, where the LLMs are known to have hallucination issues, we first disentangle **Understanding Literature** into two atomic abilities, i) **Understanding** the unstructured text from research papers by performing scientific claim verification, and ii) Ability to interact with structured Knowledge-Graph Question-Answering~(KGQA) as a form of **Literature** grounding. We then formulate a novel agent task, dubbed KGCheck, using KGQA and domain-based Retrieval-Augmented Generation (RAG) to identify the factual errors of existing large-scale knowledge graph databases. We collect over two thousand data for two atomic tasks and 225 high-quality annotated data for the agent task. Surprisingly, we discover that state-of-the-art agents, both daily scenarios and biomedical ones, have either failed or inferior performance on our benchmark. We then introduce a simple yet effective baseline, dubbed `BKGAgent`. On the widely used popular dataset, we discover over 90 factual errors which yield the effectiveness of our approach, yields substantial value for both the research community or practitioners in the biomedical domain.
15
+
16
+ ## Overview
17
+ <details open>
18
+ <summary>Tasks</summary>
19
+
20
+ * **KGCheck**: Given a knowledge graph and a scientific claim, the agent needs to check whether the claim is supported by the knowledge graph. The agent can interact with the knowledge graph by asking questions and receiving answers.
21
+ * **KGQA**: Given a knowledge graph and a question, the agent needs to answer the question based on the knowledge graph.
22
+ * **SCV**: Given a scientific claim and a research paper, the agent needs to check whether the claim is supported by the research paper.
23
+
24
+ </details>
25
+
26
+ <details open>
27
+ <summary>Code Structure</summary>
28
+ </details>
29
+
30
+ <details open>
31
+ <summary>Baseline</summary>
32
+ </details>
33
+
34
+ <details open>
35
+ <summary>Dataset</summary>
36
+ </details>
37
+
38
+ ## News and Updates
39
+ [2024-06-06] `BioKGBench` v0.1.0 is released.
40
+
41
+ ## Installation
42
+ This project has provided an environment setting file of conda, users can easily reproduce the environment by the following commands:
43
+ ```bash
44
+ conda create -n agent4s-biokg python=3.10
45
+ conda activate agent4s-biokg
46
+ pip install -r requirements.txt
47
+ ```
48
+
49
+ ## Getting Started
50
+
51
+ **Obtaining dataset**:
52
+ The dataset can be found in the [release]. The dataset is divided into three parts: `KGCheck`, `KGQA`, and `SCV`, every part is split into `Dev` and `Test`.
53
+
54
+ **Running Baseline**:
55
+
56
+ * `KGCheck`:
57
+ ```bash
58
+ ```
59
+ * `KGQA`:
60
+ ```bash
61
+ ```
62
+ * `SCV`:
63
+ ```bash
64
+ ```
65
+
66
+ ## Acknowledgement
67
+
68
+ `BioKGBench` is an open-source project for Agent evaluation created by researchers in **Westlake Auto Lab** and **CAIRI Lab**. We encourage researchers interested in LLM Agent and other related fields to contribute to this project!
69
+
70
+ ## Citation
71
+
72
+ ## Contact
73
+ For adding new features, looking for helps, or reporting bugs associated with `BioKGBench`, please open a [GitHub issue](https://github.com/A4Bio/ProteinInvBench/issues) and [pull request](https://github.com/A4Bio/ProteinInvBench/pulls) with the tag `new features`, `help wanted`, or `enhancement`. Feel free to contact us through email if you have any questions.
74
+
75
+ - Xinna Lin([email protected]), Westlake University
76
+ - Siqi Ma([email protected]), Westlake University
77
+ - Junjie Shan([email protected]), Westlake University
78
+ - Xiaojing Zhang([email protected]), Westlake University
79
+
80
+ ## TODO
81
+
82
+ 1. Update dataset
83
+ 2. Support pip installation