LLAMA-QA-AudioFiles / llm_ops.py
Samarth991's picture
adding notebbok and debugging main
c88baef
raw
history blame contribute delete
No virus
1.5 kB
import os
import torch
from langchain import HuggingFacePipeline
from transformers import AutoTokenizer
import transformers
from langchain import HuggingFaceHub
def get_openai_chat_model(API_key):
try:
from langchain.llms import OpenAI
except ImportError as err:
raise "{}, unable to load openAI. Please install openai and add OPENAIAPI_KEY"
os.environ["OPENAI_API_KEY"] = API_key
llm = OpenAI()
return llm
def get_llama_model(temperature=0,api_key=None,max_tokens=2048):
model_id = "meta-llama/Llama-2-7b-chat-hf"
llm = None
try:
tokenizer = AutoTokenizer.from_pretrained(model_id)
pipeline = transformers.pipeline("text-generation",
model=model_id,
tokenizer=tokenizer,
torch_dtype=torch.bfloat16,
trust_remote_code=True,
device_map="auto",
max_length=1000,
eos_token_id=tokenizer.eos_token_id
)
llm = HuggingFacePipeline(pipeline = pipeline, model_kwargs = {'temperature':temperature})
except:
raise "User not autorized to access the Model"
return llm
def get_model_from_hub(api_key,temperature=0.1,max_tokens=2048,model_id="meta-llama/Llama-2-7b-chat-hf"):
llm = HuggingFaceHub(huggingfacehub_api_token=api_key,
repo_id=model_id,
model_kwargs={"temperature": temperature, "max_new_tokens": max_tokens})
return llm