Skip to content

EncryptedBinary/Meta_analysis

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

35 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Empowering Meta-Analysis: Leveraging Large Language Models for Scientific Synthesis

Official Implementation for IEEE BigData 2024

Abstract

This study investigates automating meta-analysis in scientific documents using large language models (LLMs). Meta-analysis synthesizes findings from multiple studies but is labor-intensive when done manually. Our approach fine-tunes LLMs for efficient, automated meta-analysis using Retrieval Augmented Generation (RAG) and a novel loss metric, Inverse Cosine Distance (ICD). The fine-tuned models achieved 87.6% relevant meta-analysis abstracts and reduced irrelevance from 4.56% to 1.9%, demonstrating efficiency in a low-resource environment.

Here is the available Meta-Analysis Dataset (MAD)


Dataset Statistics

Metric Actual Chunked
Types of Domains : Scientific Studies
Min. input (Sj) context length 733 1005
Max. input (Sj) context length 32,767 2,000
Avg. input (Sj) context length 16,890.22 1,542.32
Min. labels (yj) context length 104 104
Max. labels (yj) context length 2,492 2,492
Avg. labels (yj) context length 1,446.45 1,446.45
Total Instances 625 7,447
Human Evaluators DetailsCount
Total no. of evaluators13
No. of female evaluators4
No. of male evaluators9
Avg. age23
ProfessionStudent, Engineer

Methodology

  1. Data Preprocessing: Chunking datasets for LLM context length restrictions.
  2. Fine-Tuning: LLMs fine-tuned using Inverse Cosine Distance (ICD) to maximize relevance.
  3. RAG Integration: Semantic search matches context with relevant data chunks for summary generation.
  4. Evaluation: Human evaluators assess abstract readability and relevance.

image

Fig. 1: (a) Paraphraser-based approach [1], [2]; (b) Retrieval Augmentation Generation [3], [4]; (c) Our novel approach with fine-tuned LLMs.

Result

Model Performance on Summarization Quality Across Benchmark Datasets

This table compares model performance on benchmark datasets for summarization quality without fine-tuning, enabling assessment across varying context lengths.

Method Models Open-i (BLEU ↑) Open-i (ROUGE ↑) writer_summaries (BLEU ↑) writer_summaries (ROUGE ↑) CL-SciSumm (BLEU ↑) CL-SciSumm (ROUGE ↑)
Established
GPT-4 with ICL 46.0 68.2 - - - -
InstructGPT davinci v2 - - - - 48 -
GCN Hybrid - - - - - 33.88
Context length restricted LLMs
Pre-trained Falcon 7B 0.19 3.17 0.76 5.19 0.71 2.21
Pre-trained Gemma 7B 2.13 8.81 4.47 30.28 2.44 20.78
Pre-trained Orca-2 7B 3.53 8.36 4.29 22.51 2.86 15.55
Pre-trained StableLM-Base-Alpha 7B 2.01 2.45 3.56 15.36 1.17 16.58
Pre-trained Llama-2 7B 4.81 10.28 5.21 31.61 3.01 22.84
Pre-trained Mistral-v0.1 7B 1.21 6.57 1.62 6.37 0.36 2.55
Ours Llama-2 7B FT 10.14 27.39 12.66 31.36 7.15 25.22
Ours Mistral-v0.1 7B FT 12.42 31.57 14.56 35.56 8.38 27.29

Notes

  1. Open-i: Medical radiological dataset. Generated summaries from 100 samples.
  2. writer_summaries: Article summarization dataset, evaluated on 120 samples.
  3. CL-SciSumm: Large corpus containing scientific article data, evaluated on 20 samples. Chunking required due to context length limitations.
  4. Established: Pre-established methods from the cited papers for the three specific datasets. BLEU and ROUGE scores are not comparable with the other models due to different evaluation methodologies.

🚀 Quick Start

To get started with our models, follow the steps below.

1. Clone the Repository

git clone https://github.com/EncryptedBinary/Meta_analysis.git
cd Meta_analysis
  1. Install Required Packages Run the following commands to install the necessary libraries:
!pip install transformers trl accelerate torch bitsandbytes peft datasets -qU
!pip install langchain

🧪 Train-Test-Split

  • Training: 400 meta-analysis documents
  • Validation: 75 meta-analysis documents
  • Testing: 50 meta-analysis documents

Feel free to modify the splits or experiment with different datasets based on your use case.

📚 Model Training

For those interested in fine-tuning the models further, we recommend checking out the train.py script, which includes hyperparameters and configurations for:

  • Epochs: 10
  • Loss Function: Inverse Cosine Distance (ICD)
  • Optimization: Using bitsandbytes for efficient scaling

References


About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published