GLM-Edge-1.5B-Chat
中文阅读, 点击这里
Inference with Transformers
Installation
Install the transformers library from the source code:
pip install git+https://github.com/huggingface/transformers.git
Inference
from transformersimport AutoModelForCausalLM, AutoTokenizerMODEL_PATH ="THUDM/glm-edge-1.5b-chat"tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)model = AutoModelForCausalLM.from_pretrained(MODEL_PATH, device_map="auto")message = [{"role":"user","content":"hello!"}]inputs = tokenizer.apply_chat_template( message, return_tensors="pt", add_generation_prompt=True, return_dict=True,).to(model.device)generate_kwargs = {"input_ids": inputs["input_ids"],"attention_mask": inputs["attention_mask"],"max_new_tokens":128,"do_sample":False,}out = model.generate(**generate_kwargs)print(tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True))
License
The usage of this model’s weights is subject to the terms outlined in theLICENSE.
- Downloads last month
- 962
Inference ProvidersNEW
This model isn't deployed by any Inference Provider.🙋Ask for provider support
HF Inference deployability: The model authors have turned it off explicitly.