Movatterモバイル変換


[0]ホーム

URL:


Skip to main content
OurBuilding Ambient Agents with LangGraph course is now available on LangChain Academy!
Open In ColabOpen on GitHub

Self Hosted

Let's load theSelfHostedEmbeddings,SelfHostedHuggingFaceEmbeddings, andSelfHostedHuggingFaceInstructEmbeddings classes.

import runhouseas rh
from langchain_community.embeddingsimport(
SelfHostedEmbeddings,
SelfHostedHuggingFaceEmbeddings,
SelfHostedHuggingFaceInstructEmbeddings,
)
# For an on-demand A100 with GCP, Azure, or Lambda
gpu= rh.cluster(name="rh-a10x", instance_type="A100:1", use_spot=False)

# For an on-demand A10G with AWS (no single A100s on AWS)
# gpu = rh.cluster(name='rh-a10x', instance_type='g5.2xlarge', provider='aws')

# For an existing cluster
# gpu = rh.cluster(ips=['<ip of the cluster>'],
# ssh_creds={'ssh_user': '...', 'ssh_private_key':'<path_to_key>'},
# name='my-cluster')
embeddings= SelfHostedHuggingFaceEmbeddings(hardware=gpu)
text="This is a test document."
query_result= embeddings.embed_query(text)

And similarly for SelfHostedHuggingFaceInstructEmbeddings:

embeddings= SelfHostedHuggingFaceInstructEmbeddings(hardware=gpu)

Now let's load an embedding model with a custom load function:

defget_pipeline():
from transformersimport(
AutoModelForCausalLM,
AutoTokenizer,
pipeline,
)

model_id="facebook/bart-base"
tokenizer= AutoTokenizer.from_pretrained(model_id)
model= AutoModelForCausalLM.from_pretrained(model_id)
return pipeline("feature-extraction", model=model, tokenizer=tokenizer)


definference_fn(pipeline, prompt):
# Return last hidden state of the model
ifisinstance(prompt,list):
return[emb[0][-1]for embin pipeline(prompt)]
return pipeline(prompt)[0][-1]
embeddings= SelfHostedEmbeddings(
model_load_fn=get_pipeline,
hardware=gpu,
model_reqs=["./","torch","transformers"],
inference_fn=inference_fn,
)
query_result= embeddings.embed_query(text)

Related


[8]ページ先頭

©2009-2025 Movatter.jp