|
|
|
|
| from transformers import AutoTokenizer |
| from vllm import LLM, SamplingParams |
| from arguments import get_args |
| from dataset import load_data, get_inputs |
| import torch |
| import os |
|
|
| def get_prompt_list(args): |
|
|
| |
| tokenizer = AutoTokenizer.from_pretrained(args.model_id) |
|
|
| |
| if args.eval_dataset == "doc2dial": |
| input_datapath = os.path.join(args.data_folder, args.doc2dial_path) |
| elif args.eval_dataset == "convfinqa": |
| input_datapath = os.path.join(args.data_folder, args.convfinqa_path) |
| elif args.eval_dataset == "quac": |
| input_datapath = os.path.join(args.data_folder, args.quac_path) |
| elif args.eval_dataset == "qrecc": |
| input_datapath = os.path.join(args.data_folder, args.qrecc_path) |
| elif args.eval_dataset == "doqa_cooking": |
| input_datapath = os.path.join(args.data_folder, args.doqa_cooking_path) |
| elif args.eval_dataset == "doqa_travel": |
| input_datapath = os.path.join(args.data_folder, args.doqa_travel_path) |
| elif args.eval_dataset == "doqa_movies": |
| input_datapath = os.path.join(args.data_folder, args.doqa_movies_path) |
| elif args.eval_dataset == "coqa": |
| input_datapath = os.path.join(args.data_folder, args.coqa_path) |
| elif args.eval_dataset == "sqa": |
| input_datapath = os.path.join(args.data_folder, args.sqa_path) |
| elif args.eval_dataset == "topiocqa": |
| input_datapath = os.path.join(args.data_folder, args.topiocqa_path) |
| elif args.eval_dataset == "inscit": |
| input_datapath = os.path.join(args.data_folder, args.inscit_path) |
| elif args.eval_dataset == "hybridial": |
| input_datapath = os.path.join(args.data_folder, args.hybridial_path) |
|
|
| else: |
| raise Exception("please input a correct eval_dataset name!") |
| |
| data_list = load_data(input_datapath) |
| print("number of samples in the dataset:", len(data_list)) |
| prompt_list = get_inputs(data_list, args.eval_dataset, tokenizer, num_ctx=args.num_ctx, max_output_len=args.out_seq_len) |
|
|
| return prompt_list |
|
|
|
|
| def main(): |
| args = get_args() |
| |
| |
| bos_token = "<|begin_of_text|>" |
|
|
| |
| model_path = os.path.join(args.model_folder, args.model_name) |
| |
| |
| prompt_list = get_prompt_list(args) |
| |
| |
| output_datapath = os.path.join(args.output_folder, "%s_output.txt" % args.eval_dataset) |
|
|
| |
| sampling_params = SamplingParams(temperature=0, top_k=1, max_tokens=args.max_tokens) |
|
|
| |
| model_vllm = LLM(model_path, tensor_parallel_size=8) |
|
|
| output_list = [] |
| for prompt in prompt_list: |
| prompt = bos_token + prompt |
| output = model_vllm.generate([prompt], sampling_params)[0] |
| generated_text = output.outputs[0].text |
| generated_text = generated_text.strip().replace("\n", " ") |
| |
| |
| output_list.append(generated_text) |
|
|
| print("writing to %s" % output_datapath) |
| with open(output_datapath, "w") as f: |
| for output in output_list: |
| f.write(output + "\n") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|