| | --- |
| | library_name: transformers |
| | license: apache-2.0 |
| | language: |
| | - en |
| | --- |
| | |
| | ### Example |
| | ``` |
| | from transformers import AutoModelForCausalLM, AutoTokenizer |
| | model_name = "jaeyong2/QuerySense-Preview" |
| | model = AutoModelForCausalLM.from_pretrained(model_name) |
| | tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | |
| | prompt = """ |
| | # Role |
| | You are an AI that receives Questions and Context from users as input and preprocesses the Questions. |
| | |
| | # Instruction |
| | - If the user's Questions contains enough information to create an answer, use the user's Questions as is. |
| | - If the information is insufficient or the Context is insufficient, please rephrase the Questions with the necessary information. |
| | - If there is insufficient information to generate an answer and there is no Context, it will automatically fill in the appropriate information. |
| | |
| | # input |
| | - Context : Previous conversations or related Context or related information entered by the user (Optional) |
| | - Question : User's Questions (Required) |
| | """.strip() |
| | |
| | content =""" |
| | Context : |
| | Question : name |
| | """.strip() |
| | |
| | system = {"role":"system", "content":prompt} |
| | user = {"role":"user", "content":content} |
| | messages = [system, user] |
| | text = tokenizer.apply_chat_template( |
| | messages, |
| | tokenize=False, |
| | add_generation_prompt=True, |
| | enable_thinking=False # Switches between thinking and non-thinking modes. Default is True. |
| | ) |
| | model_inputs = tokenizer([text], return_tensors="pt").to(model.device) |
| | |
| | # conduct text completion |
| | generated_ids = model.generate( |
| | **model_inputs, |
| | max_new_tokens=32768 |
| | ) |
| | output_ids = generated_ids[0][len(model_inputs.input_ids[0]):].tolist() |
| | content = tokenizer.decode(output_ids, skip_special_tokens=True).strip("\n") |
| | |
| | print("content:", content) |
| | ``` |
| |
|
| | ### result |
| | ``` |
| | content: what is the name of the product? |
| | ``` |
| |
|
| |
|
| | ## License |
| | - Qwen/Qwen3-1.7B : https://huggingface.co/Qwen/Qwen3-1.7B/blob/main/LICENSE |
| |
|
| |
|
| | ## Acknowledgement |
| | This research is supported by **TPU Research Cloud program**. |