diff --git a/trustgraph-flow/trustgraph/model/text_completion/tgi/llm.py b/trustgraph-flow/trustgraph/model/text_completion/tgi/llm.py index 67371726..fa7c15c0 100755 --- a/trustgraph-flow/trustgraph/model/text_completion/tgi/llm.py +++ b/trustgraph-flow/trustgraph/model/text_completion/tgi/llm.py @@ -77,16 +77,10 @@ class Processor(LlmService): json=request, ) as response: - print("GOT A SRESPONSE") - if response.status != 200: raise RuntimeError("Bad status: " + str(response.status)) - print("GOT A GOOD STATUS") - resp = await response.json() - - print("RESPONSE>", resp) inputtokens = resp["usage"]["prompt_tokens"] outputtokens = resp["usage"]["completion_tokens"]