diff --git a/concordia/language_model/together_ai.py b/concordia/language_model/together_ai.py index b5e94fa..03f80f3 100644 --- a/concordia/language_model/together_ai.py +++ b/concordia/language_model/together_ai.py @@ -154,6 +154,7 @@ def sample_choice( ) -> tuple[int, str, dict[str, float]]: def _sample_choice(response: str) -> float: + augmented_prompt = prompt + response messages = [ { 'role': 'system', @@ -176,7 +177,7 @@ def _sample_choice(response: str) -> float: ), }, {'role': 'assistant', 'content': 'sleeping.'}, - {'role': 'user', 'content': prompt + response}, + {'role': 'user', 'content': augmented_prompt}, ] result = None @@ -198,6 +199,7 @@ def _sample_choice(response: str) -> float: except together.error.RateLimitError as err: if attempts >= _NUM_SILENT_ATTEMPTS: print(f' Exception: {err}') + print(f' Exception prompt: {augmented_prompt}') continue else: break @@ -205,7 +207,8 @@ def _sample_choice(response: str) -> float: if result: lp = sum(result.choices[0].logprobs.token_logprobs) else: - raise ValueError('Failed to get logprobs.') + raise ValueError( + f'Failed to get logprobs.\nException prompt: {augmented_prompt}') return lp