diff --git a/graph_of_thoughts/language_models/abstract_language_model.py b/graph_of_thoughts/language_models/abstract_language_model.py index cead63c..e39f322 100644 --- a/graph_of_thoughts/language_models/abstract_language_model.py +++ b/graph_of_thoughts/language_models/abstract_language_model.py @@ -36,7 +36,7 @@ class AbstractLanguageModel(ABC): self.model_name: str = model_name self.cache = cache if self.cache: - self.respone_cache: Dict[str, List[Any]] = {} + self.response_cache: Dict[str, List[Any]] = {} self.load_config(config_path) self.prompt_tokens: int = 0 self.completion_tokens: int = 0 @@ -63,7 +63,7 @@ class AbstractLanguageModel(ABC): """ Clear the response cache. """ - self.respone_cache.clear() + self.response_cache.clear() @abstractmethod def query(self, query: str, num_responses: int = 1) -> Any: diff --git a/graph_of_thoughts/language_models/chatgpt.py b/graph_of_thoughts/language_models/chatgpt.py index 4f63d61..52dc1d3 100644 --- a/graph_of_thoughts/language_models/chatgpt.py +++ b/graph_of_thoughts/language_models/chatgpt.py @@ -73,8 +73,8 @@ class ChatGPT(AbstractLanguageModel): :return: Response(s) from the OpenAI model. :rtype: Dict """ - if self.cache and query in self.respone_cache: - return self.respone_cache[query] + if self.cache and query in self.response_cache: + return self.response_cache[query] if num_responses == 1: response = self.chat([{"role": "user", "content": query}], num_responses) @@ -98,7 +98,7 @@ class ChatGPT(AbstractLanguageModel): total_num_attempts -= 1 if self.cache: - self.respone_cache[query] = response + self.response_cache[query] = response return response @backoff.on_exception(backoff.expo, OpenAIError, max_time=10, max_tries=6) diff --git a/graph_of_thoughts/language_models/llamachat_hf.py b/graph_of_thoughts/language_models/llamachat_hf.py index d423a50..cc413ab 100644 --- a/graph_of_thoughts/language_models/llamachat_hf.py +++ b/graph_of_thoughts/language_models/llamachat_hf.py @@ -84,8 +84,8 @@ class Llama2HF(AbstractLanguageModel): :return: Response(s) from the LLaMA 2 model. :rtype: List[Dict] """ - if self.cache and query in self.respone_cache: - return self.respone_cache[query] + if self.cache and query in self.response_cache: + return self.response_cache[query] sequences = [] query = f"<>You are a helpful assistant. Always follow the intstructions precisely and output the response exactly in the requested format.<>\n\n[INST] {query} [/INST]" for _ in range(num_responses): @@ -104,7 +104,7 @@ class Llama2HF(AbstractLanguageModel): for sequence in sequences ] if self.cache: - self.respone_cache[query] = response + self.response_cache[query] = response return response def get_response_texts(self, query_responses: List[Dict]) -> List[str]: