minor example fix (#423)
This commit is contained in:
		
							parent
							
								
									d3912c70cf
								
							
						
					
					
						commit
						3b8e519f77
					
				|  | @ -22,6 +22,7 @@ elif chosen_inference == "ollama": | |||
|         model_id="ollama_chat/llama3.2", | ||||
|         api_base="http://localhost:11434",  # replace with remote open-ai compatible server if necessary | ||||
|         api_key="your-api-key",  # replace with API key if necessary | ||||
|         num_ctx=8192 # ollama default is 2048 which will often fail horribly. 8192 works for easy tasks, more is better. Check https://huggingface.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator to calculate how much VRAM this will need for the selected model. | ||||
|     ) | ||||
| 
 | ||||
| elif chosen_inference == "litellm": | ||||
|  | @ -48,4 +49,4 @@ print("ToolCallingAgent:", agent.run("What's the weather like in Paris?")) | |||
| 
 | ||||
| agent = CodeAgent(tools=[get_weather], model=model) | ||||
| 
 | ||||
| print("ToolCallingAgent:", agent.run("What's the weather like in Paris?")) | ||||
| print("CodeAgent:", agent.run("What's the weather like in Paris?")) | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue