Add example to use any LLM
This commit is contained in:
parent
57b35337c2
commit
b6fc583d96
|
@ -0,0 +1,30 @@
|
|||
from agents import OpenAIEngine, AnthropicEngine, HfApiEngine, CodeAgent
|
||||
from dotenv import load_dotenv
|
||||
|
||||
load_dotenv()
|
||||
|
||||
openai_engine = OpenAIEngine(model_name="gpt-4o")
|
||||
|
||||
agent = CodeAgent([], llm_engine=openai_engine)
|
||||
|
||||
print("\n\n##############")
|
||||
print("Running OpenAI agent:")
|
||||
agent.run("What is the 10th Fibonacci Number?")
|
||||
|
||||
|
||||
anthropic_engine = AnthropicEngine()
|
||||
|
||||
agent = CodeAgent([], llm_engine=anthropic_engine)
|
||||
|
||||
print("\n\n##############")
|
||||
print("Running Anthropic agent:")
|
||||
agent.run("What is the 10th Fibonacci Number?")
|
||||
|
||||
# Here, our token stored as HF_TOKEN environment variable has accesses 'Make calls to the serverless Inference API' and 'Read access to contents of all public gated repos you can access'
|
||||
llama_engine = HfApiEngine(model="meta-llama/Llama-3.3-70B-Instruct")
|
||||
|
||||
agent = CodeAgent([], llm_engine=llama_engine)
|
||||
|
||||
print("\n\n##############")
|
||||
print("Running Llama3.3-70B agent:")
|
||||
agent.run("What is the 10th Fibonacci Number?")
|
Loading…
Reference in New Issue