Movatterモバイル変換


[0]ホーム

URL:


Hugging Face's logoHugging Face

GLM-Edge-1.5B-Chat

中文阅读, 点击这里

Inference with Transformers

Installation

Install the transformers library from the source code:

pip install git+https://github.com/huggingface/transformers.git

Inference

from transformersimport AutoModelForCausalLM, AutoTokenizerMODEL_PATH ="THUDM/glm-edge-1.5b-chat"tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)model = AutoModelForCausalLM.from_pretrained(MODEL_PATH, device_map="auto")message = [{"role":"user","content":"hello!"}]inputs = tokenizer.apply_chat_template(    message,    return_tensors="pt",    add_generation_prompt=True,    return_dict=True,).to(model.device)generate_kwargs = {"input_ids": inputs["input_ids"],"attention_mask": inputs["attention_mask"],"max_new_tokens":128,"do_sample":False,}out = model.generate(**generate_kwargs)print(tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True))

License

The usage of this model’s weights is subject to the terms outlined in theLICENSE.

Downloads last month
962
Safetensors
Model size
1.59B params
Tensor type
BF16
·
Inference ProvidersNEW
This model isn't deployed by any Inference Provider.🙋Ask for provider support
HF Inference deployability: The model authors have turned it off explicitly.

Model tree forTHUDM/glm-edge-1.5b-chat

Finetunes
1 model
Quantizations
5 models

Spaces usingTHUDM/glm-edge-1.5b-chat2

Collection includingTHUDM/glm-edge-1.5b-chat


[8]ページ先頭

©2009-2025 Movatter.jp