Training procedure
Framework versions
PEFT 0.4.0
PEFT 0.4.0
use
in https://github.com/hiyouga/ChatGLM-Efficient-Tuning/tree/main
CUDA_VISIBLE_DEVICES=3 nohup python src/web_demo.py \
--model_name_or_path /HOME/jack/model/chatglm-6b \
--checkpoint_dir paper_meta\ \
> log_web_demo.txt 2>&1 & tail -f log_web_demo.txt
🚩Citation
Please cite the following paper if you use jackkuo/ChatPaperGPT_32k in your work.
@INPROCEEDINGS{10412837,
author={Guo, Menghao and Wu, Fan and Jiang, Jinling and Yan, Xiaoran and Chen, Guangyong and Li, Wenhui and Zhao, Yunhong and Sun, Zeyi},
booktitle={2023 IEEE International Conference on Knowledge Graph (ICKG)},
title={Investigations on Scientific Literature Meta Information Extraction Using Large Language Models},
year={2023},
volume={},
number={},
pages={249-254},
keywords={Measurement;Knowledge graphs;Information retrieval;Data mining;Task analysis;information extraction;large language model;scientific literature},
doi={10.1109/ICKG59574.2023.00036}}
- Downloads last month
- 3