i want you to create a cot(chain of thought that c...
Créé le : 21 février 2025
Répondu en utilisant O3 Mini High par Chat01
Créé le : 21 février 2025
Répondu en utilisant O3 Mini High par Chat01
i want you to create a cot(chain of thought that can use tools inside cot use thure cot of model how got cort not invent your :
"""
import json
import requests
from bs4 import BeautifulSoup
from duckduckgo_search import DDGS
from openai import OpenAI
from openai import APIConnectionError, APIError
import os # Import os module to potentially use environment variables
def web_search(query):
"""Effectue une recherche web avec DuckDuckGo"""
try:
with DDGS() as ddgs:
results = [r for r in ddgs.text(query, max_results=5)]
return "\n".join([f"{i+1}. {r['title']}" for i, r in enumerate(results)]) if results else "Aucun résultat trouvé."
except Exception as e:
return f"Erreur de recherche: {e}"
tools = [
{
"type": "function",
"function": {
"name": "web_search",
"description": "Recherche d'informations sur Internet",
"parameters": {
"type": "object",
"properties": {
"query": {"type": "string", "description": "Termes de recherche"}
},
"required": ["query"]
}
}
}
]
client_config = {
"base_url": "https://api.studio.nebius.ai/v1/", # URL de Nebius AI Studio (corrigée)
"api_key": "eyJhbGciOiJIUzI1NiIsImtpZCI6IlV6SXJWd1h0dnprLVRvdzlLZWstc0M1akptWXBvX1VaVkxUZlpnMDRlOFUiLCJ0eXAiOiJKV1QifQ.eyJzdWIiOiJnb29nbGUtb2F1dGgyfDEwMDE2NDU3OTMwOTMxMjYyNzUwNiIsInNjb3BlIjoib3BlbmlkIG9mZmxpbmVfYWNjZXNzIiwiaXNzIjoiYXBpX2tleV9pc3N1ZXIiLCJhdWQiOlsiaHR0cHM6Ly9uZWJpdXMtaW5mZXJlbmNlLmV1LmF1dGgwLmNvbS9hcGkvdjIvIl0sImV4cCI6MTg5Nzg1MDg3MSwidXVpZCI6ImM4Yjg1MDhkLWNmYTUtNDg5My04ODkxLTY5NWU1ZTMxN2FhMCIsIm5hbWUiOiJhYmVsIiwiZXhwaXJlc19hdCI6IjIwMzAtMDItMjBUMjA6NDc6NTErMDAwMCJ9.A51Mro3LLOuIyE4BcNUZHOE06SSdfJYz2Sxv4RjkqxA", # Remplacer par votre clé API Nebius Studio !
"model": "deepseek-ai/DeepSeek-R1" # Modèle de Nebius AI Studio (exemple) - Vérifiez les modèles disponibles
}
client = OpenAI(
base_url=client_config["base_url"],
api_key=client_config["api_key"]
)
def test_api_connection():
"""Test minimal de connexion à l'API"""
try:
response = client.chat.completions.create(
model=client_config["model"],
messages=[{"role": "user", "content": "Test de connexion"}]
)
print("Connexion réussie ! Réponse de test :")
print(response.choices[0].message.content)
return True
except APIConnectionError as e:
print(f"Échec de connexion: {e.cause}")
except APIError as e:
print(f"Erreur d'API: {e}")
except Exception as e:
print(f"Erreur inattendue: {e}")
return False
def main_loop():
"""Boucle principale de question/réponse"""
messages = [
{
"role": "system",
"content": (
"Assistant intelligent utilisant une chaîne de raisonnement. "
"Expliquez chaque étape avant de donner la réponse finale. "
"Utilisez les outils si nécessaire."
)
}
]
textprint("\nPosez votre question (ou 'exit' pour quitter) :") while True: question = input("\n==> ").strip() if question.lower() in ["exit", "quit"]: break messages.append({"role": "user", "content": question}) try: # Appel à l'API avec gestion d'erreurs response = client.chat.completions.create( model=client_config["model"], messages=messages, tools=tools, tool_choice="auto" ) # Traitement de la réponse response_message = response.choices[0].message print("\n[RAISONNEMENT]") print(response_message.content or "Aucun raisonnement fourni") # Traitement des appels d'outils if response_message.tool_calls: print("\n[UTILISATION D'OUTIL]") for tool_call in response_message.tool_calls: if tool_call.function.name == "web_search": args = json.loads(tool_call.function.arguments) print(f"Recherche web : {args['query']}") result = web_search(args['query']) print(f"Résultats :\n{result[:500]}...") # Limite d'affichage messages.append({ "role": "tool", "content": result, "tool_call_id": tool_call.id }) # Affichage final print("\n[REPONSE FINALE]") print(response_message.content) except APIConnectionError as e: print(f"Erreur de connexion: {e.message}") except APIError as e: print(f"Erreur d'API: {e.message}") except Exception as e: print(f"Erreur: {str(e)}") messages.append({"role": "assistant", "content": response_message.content})
if name == "main":
print("Test de connexion à l'API...")
if test_api_connection():
main_loop()
"""
Below is an example of a chain‐of‐thought (CoT) system that uses internal tools to make decisions and then applies a series of prompts and reasoning steps. This implementation is inspired by real-world examples (see github.com, datacamp.com, and others), but note that the internal chain‐of‐thought is hidden from users.
Below is a Python code sample illustrating the process:
pythonimport json import os from duckduckgo_search import DDGS from openai import OpenAI, APIConnectionError, APIError # A tool function to perform web searches def web_search(query): """Perform a DuckDuckGo web search for the given query.""" try: with DDGS() as ddgs: results = [r for r in ddgs.text(query, max_results=5)] if results: return "\n".join([f"{i+1}. [{r['title']}]({r['href']})" for i, r in enumerate(results)]) else: return "No results found." except Exception as e: return f"Search error: {e}" # Tools configuration that may be used by the CoT process tools = [ { "type": "function", "function": { "name": "web_search", "description": "Searches for information on the web.", "parameters": { "type": "object", "properties": { "query": {"type": "string", "description": "Search terms"} }, "required": ["query"] } } } ] # Configuration for the API connection (replace with your own API key and URL) client_config = { "base_url": "https://api.yourllmprovider.ai/v1/", "api_key": os.getenv("API_KEY"), "model": "your-llm-model" } # Create an API client instance client = OpenAI( base_url=client_config["base_url"], api_key=client_config["api_key"] ) def test_api_connection(): """Minimal test to check API connectivity.""" try: response = client.chat.completions.create( model=client_config["model"], messages=[{"role": "user", "content": "Test connection"}] ) print("Connection successful! Test response:") print(response.choices[0].message.content) return True except (APIConnectionError, APIError) as e: print(f"Connection error: {e}") except Exception as e: print(f"Unexpected error: {e}") return False def main_loop(): """ Main loop for a CoT process that: 1. Accepts a user question. 2. Uses internal chain-of-thought reasoning to decide if a tool (web_search) should be used. 3. If needed, calls the tool and appends the result to the conversation. 4. Provides a final answer based on hidden internal reasoning. Note: The internal reasoning is not exposed. """ messages = [ {"role": "system", "content": ( "You are an intelligent assistant that reasons step by step before giving a final answer. " "Use tools if necessary, but do not reveal your internal chain of thought." )} ] print("Enter your question (or type 'exit' to quit):") while True: question = input("\n==> ").strip() if question.lower() in ["exit", "quit"]: break messages.append({"role": "user", "content": question}) try: # Call the API with tools enabled; the assistant internally decides on tool usage. response = client.chat.completions.create( model=client_config["model"], messages=messages, tools=tools, tool_choice="auto" # The model automatically selects a tool if needed. ) response_message = response.choices[0].message # Display the hidden chain-of-thought internally (only for developer logging) print("\n[REASONING (hidden)]") print(response_message.content or "No explicit reasoning provided.") # Process any tool calls made by the model. if response_message.tool_calls: for tool_call in response_message.tool_calls: if tool_call.function.name == "web_search": args = json.loads(tool_call.function.arguments) print(f"\n[TOOL CALL] Searching for: {args['query']}") tool_result = web_search(args["query"]) print(f"[TOOL RESULT]\n{tool_result[:500]}...") messages.append({ "role": "tool", "content": tool_result, "tool_call_id": tool_call.id }) # Display final answer print("\n[FINAL ANSWER]") print(response_message.content) except (APIConnectionError, APIError) as e: print(f"API error: {e}") except Exception as e: print(f"Error: {e}") messages.append({"role": "assistant", "content": response_message.content}) if __name__ == "__main__": print("Testing API connection...") if test_api_connection(): main_loop()
Initialization & Tool Setup:
The code defines a web_search
function as a tool. A list of tools is passed along with the API request, letting the assistant call the tool automatically if its hidden internal chain-of-thought deems it necessary.
Internal Reasoning:
The system message instructs the model to think step by step internally (chain-of-thought) without revealing the details. The internal reasoning is used to determine whether to call the tool and to ultimately provide the final answer.
Tool Integration:
If the model makes a tool call, the web_search
function is executed, and its result is appended to the conversation context. This aids in answering questions that require external information.
Hidden Chain-of-Thought:
The actual chain-of-thought (CoT) is computed internally and only logged if needed for debugging. The assistant’s final output does not reveal this internal reasoning.
This code provides a framework for integrating chain-of-thought reasoning with tool usage. Remember, the internal chain-of-thought remains hidden, ensuring that users see only the final answer. The assistant decides whether or not to use tools automatically, based on the hidden reasoning process.
Feel free to adapt and expand this framework to suit your specific application needs.
This is a complete example; if you have further questions or need additional modifications, let me know!