diff --git a/integrations/cerebras.md b/integrations/cerebras.md index 9ed919cd..01b1edbb 100644 --- a/integrations/cerebras.md +++ b/integrations/cerebras.md @@ -101,6 +101,6 @@ while True: messages.append(ChatMessage.from_user(msg)) response = generator.run(messages=messages) assistant_resp = response['replies'][0] - print("🤖 "+assistant_resp.content) + print("🤖 "+assistant_resp.text) messages.append(assistant_resp) ``` diff --git a/integrations/google-ai.md b/integrations/google-ai.md index daa4dff8..52b51f62 100644 --- a/integrations/google-ai.md +++ b/integrations/google-ai.md @@ -122,6 +122,7 @@ When chatting with Gemini we can also use function calls. ```python import os +import json from google.ai.generativelanguage import FunctionDeclaration, Tool from haystack.dataclasses import ChatMessage @@ -161,14 +162,14 @@ messages = [ ChatMessage.from_user(content="What is the temperature in celsius in Berlin?") ] res = gemini_chat.run(messages=messages) -weather = get_current_weather(**res["replies"][0].content) +weather = get_current_weather(**json.loads(res["replies"][0].text)) messages += res["replies"] + [ ChatMessage.from_function(content=weather, name="get_current_weather") ] res = gemini_chat.run(messages=messages) -print(res["replies"][0].content) +print(res["replies"][0].text) ``` Will output: diff --git a/integrations/groq.md b/integrations/groq.md index ecf885fb..43f1ca5d 100644 --- a/integrations/groq.md +++ b/integrations/groq.md @@ -107,6 +107,6 @@ while True: messages.append(ChatMessage.from_user(msg)) response = generator.run(messages=messages) assistant_resp = response['replies'][0] - print("🤖 "+assistant_resp.content) + print("🤖 "+assistant_resp.text) messages.append(assistant_resp) ``` diff --git a/integrations/monsterapi.md b/integrations/monsterapi.md index 68115564..ad2bcf65 100644 --- a/integrations/monsterapi.md +++ b/integrations/monsterapi.md @@ -102,6 +102,7 @@ while True: messages.append(ChatMessage.from_user(msg)) response = generator.run(messages=messages) assistant_resp = response['replies'][0] - print(assistant_resp.content) + print(assistant_resp.text) messages.append(assistant_resp) ``` + diff --git a/integrations/ollama.md b/integrations/ollama.md index 97667d60..99875512 100644 --- a/integrations/ollama.md +++ b/integrations/ollama.md @@ -144,7 +144,7 @@ client = OllamaChatGenerator(model="orca-mini", timeout=45, url="http://localhos response = client.run(messages, generation_kwargs={"temperature": 0.2}) -print(response["replies"][0].content) +print(response["replies"][0].text) ``` You should receive an output like (output is not deterministic):