make compatible with remote openai-compatible servers
This commit is contained in:
		
							parent
							
								
									95aae1ba85
								
							
						
					
					
						commit
						b3dfaddd43
					
				|  | @ -410,11 +410,13 @@ class TransformersModel(Model): | |||
| 
 | ||||
| 
 | ||||
| class LiteLLMModel(Model): | ||||
|     def __init__(self, model_id="anthropic/claude-3-5-sonnet-20240620"): | ||||
|     def __init__(self, model_id="anthropic/claude-3-5-sonnet-20240620", api_base=None, api_key=None): | ||||
|         super().__init__() | ||||
|         self.model_id = model_id | ||||
|         # IMPORTANT - Set this to TRUE to add the function to the prompt for Non OpenAI LLMs | ||||
|         litellm.add_function_to_prompt = True | ||||
|         self.api_base = api_base | ||||
|         self.api_key = api_key | ||||
| 
 | ||||
|     def __call__( | ||||
|         self, | ||||
|  | @ -432,6 +434,8 @@ class LiteLLMModel(Model): | |||
|             messages=messages, | ||||
|             stop=stop_sequences, | ||||
|             max_tokens=max_tokens, | ||||
|             api_base=self.api_base, | ||||
|             api_key=self.api_key, | ||||
|         ) | ||||
|         self.last_input_token_count = response.usage.prompt_tokens | ||||
|         self.last_output_token_count = response.usage.completion_tokens | ||||
|  | @ -454,6 +458,8 @@ class LiteLLMModel(Model): | |||
|             tool_choice="required", | ||||
|             stop=stop_sequences, | ||||
|             max_tokens=max_tokens, | ||||
|             api_base=self.api_base, | ||||
|             api_key=self.api_key, | ||||
|         ) | ||||
|         tool_calls = response.choices[0].message.tool_calls[0] | ||||
|         self.last_input_token_count = response.usage.prompt_tokens | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue