Skip to content
forked from cincynlp/FrameID

Fork of code for the paper: "Do LLMs Encode Frame Semantics? Evidence from Frame Identification"

Notifications You must be signed in to change notification settings

nd-ball/FrameID

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Code For the paper : "Do LLMs Encode Frame Semantics? Evidence from Frame Identification".

Environment Setup

  • Python: 3.12.8
  • PyTorch: 2.5.1+cu121
  • Transformers: 4.49.0

Data Formatting

cd data
python convert_csv_to_json.py --folder fn1.7/data-csv

This will generate:

json-output/

Preparing Data for Fine-Tuning

cd fine_tune
python data_for_finetuning.py --folder ../data/fn1.7/json_output --outdir finetune_data_1_7

Fine-Tuning

python finetune.py --data-dir finetune_data_1_7 --output-dir model_1_7

The trained model and adapter will be saved in:

model_1_7/

Evaluation

  1. Copy the adapter files to a new folder:
cd evaluate
mkdir lora_model_1_7
cp ../fine_tune/model_1_7/adapter_config.json lora_model_1_7/
cp ../fine_tune/model_1_7/adapter_model.safetensors lora_model_1_7/
  1. Run evaluation :
cd evaluate
python evaluate.py --dataset ../fine_tune/finetune_data_1_7/test.jsonl --adapter lora_model_1_7

To Evaluate on Downstream tasks

i. YAGS

Follow the same steps as Data Preparation steps followed for Fine-tuning and create a dataset folder yags_data,

cd evaluate
python evaluate.py --dataset yags_data/test.jsonl --adapter lora_model_1_7

ii. Artifacts

cd evaluate
python artifacts_evaluate.py --dataset artifacts_data/artifacts.jsonl --adapter lora_model_1_7

This will output evaluation results.

Contact

For questions or issues, please:

About

Fork of code for the paper: "Do LLMs Encode Frame Semantics? Evidence from Frame Identification"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%