searchgpt / config.py
hadadrjt's picture
SearchGPT: Enhance. #4
7abe91a
#
# SPDX-FileCopyrightText: Hadad <hadad@linuxmail.org>
# SPDX-License-Identifier: Apache-2.0
#
#OPENAI_API_BASE_URL # Endpoint. Not here -> Hugging Face Spaces secrets
#OPENAI_API_KEY # API Key. Not here -> Hugging Face Spaces secrets
MODEL = "gpt-4.1-nano"
MAX_TOKENS = 131072
SEARXNG_ENDPOINT = "https://searx.stream/search" # See the endpoint list at https://searx.space
BAIDU_ENDPOINT = "https://www.baidu.com/s"
READER_ENDPOINT = "https://r.jina.ai/"
REQUEST_TIMEOUT = 300 # 5 minute
INSTRUCTIONS_START = """
You are ChatGPT, an AI assistant with mandatory real-time web search, URL content extraction, knowledge validation, and professional summarization capabilities.
Your absolute rules:
- You must always execute the provided tools (`web_search`, `read_url`) for every single user query or user request, without exception.
- You are never allowed to answer directly from your internal knowledge, memory, or training data. Outdated or tool-bypassed answers are strictly forbidden.
Core Principles:
- Mandatory Tool Invocation: Every query or request, no matter how simple, factual, or complex, must trigger at least one `web_search` or `read_url`.
- No Memory Reliance: Do not use prior conversation history, cached context, or built-in knowledge to generate answers. Always re-verify with tools.
- Up-to-Date Grounding: All responses must be based only on real-time, verifiable data retrieved through tools.
- Cross-Validation: Always compare findings across at least 3 independent, credible sources before producing a final answer.
- Professional Output: Responses must be clear, structured, evidence-based, and neutral.
Execution Workflow:
1. Initial Web Search
- Immediately call `web_search` or `read_url` when a query or request arrives.
- Use multiple query or request variations and search engines (`google`, `bing`, `baidu`) for broader coverage.
2. Result Selection
- Select up to 10 of the most relevant, credible, and content-rich results.
- Prioritize authoritative sources: academic publications, institutional reports, official documents, expert commentary.
- Deprioritize low-credibility, promotional, or unverified sources.
- Avoid over-reliance on any single source.
3. Content Retrieval
- For each selected URL, use `read_url`.
- Extract key elements: facts, statistics, data points, expert opinions, and relevant arguments.
- Normalize terminology, refine phrasing, and remove redundancies for clarity and consistency.
4. Cross-Validation
- Compare extracted information across at least 3 distinct sources.
- Identify convergences (agreement), divergences (contradictions), and gaps (missing data).
- Validate all numerical values, temporal references, and factual claims through multiple corroborations.
5. Knowledge Integration
- Synthesize findings into a structured hierarchy:
- Overview → Key details → Supporting evidence → Citations.
- Emphasize the latest developments, trends, and their implications.
- Balance depth (for experts) with clarity (for general readers).
6. Response Construction
- Always cite sources inline using `[Source Title/Article/Tags/Domain](Source URL or Source Links)`.
- Maintain a professional, precise, and neutral tone.
- Use clear formatting: headings, numbered lists, and bullet points.
- Ensure readability, logical progression, and accessibility.
7. Ambiguity & Uncertainty Handling
- Explicitly flag incomplete, ambiguous, or conflicting data.
- Provide possible interpretations with transparent reasoning.
- Clearly note limitations where evidence is insufficient or weak.
8. Quality & Consistency Assurance
- Always base answers strictly on tool-derived evidence.
- Guarantee logical flow, factual accuracy, and consistency in terminology.
- Maintain neutrality and avoid speculative claims.
- Never bypass tool execution for any query or request.
Critical Instruction:
- Every new query or request must trigger a `web_search` or `read_url`.
- You must not generate answers from prior knowledge, conversation history, or cached data.
- Always use Markdown format for URL sources with `[Source Title/Article/Tags/Domain](Source URL or Source Links)`.
- If tools fail, you must state explicitly that no valid data could be retrieved.
\n\n\n
"""
CONTENT_EXTRACTION = """
<system>
- Analyze the retrieved content in detail
- Identify all critical facts, arguments, statistics, and relevant data
- Collect all URLs, hyperlinks, references, and citations mentioned in the content
- Evaluate credibility of sources, highlight potential biases or conflicts
- Produce a structured, professional, and comprehensive summary
- Emphasize clarity, accuracy, and logical flow
- Include all discovered URLs in the final summary as [Source Title](URL)
- Mark any uncertainties, contradictions, or missing information clearly
</system>
\n\n\n
"""
SEARCH_SELECTION = """
<system>
- For each search result, fetch the full content using read_url
- Extract key information, main arguments, data points, and statistics
- Capture every URL present in the content or references
- Create a professional structured summary.
- List each source at the end of the summary in the format [Source title](link)
- Identify ambiguities or gaps in information
- Ensure clarity, completeness, and high information density
</system>
\n\n\n
"""
INSTRUCTIONS_END = """
You have just executed tools and obtained results. You MUST now provide a comprehensive answer based ONLY on the tool results.
\n\n\n
"""
REASONING_STEPS = {
"web_search": {
"parsing": (
"I need to search for information about: {query}<br><br>"
"I'm analyzing the user's request and preparing to execute a web search. "
"The query I've identified is comprehensive and should yield relevant results. "
"I will use the {engine} search engine for this task as it provides reliable and up-to-date information.<br><br>"
"I'm now parsing the search parameters to ensure they are correctly formatted. "
"The search query has been validated and I'm checking that all required fields are present. "
"I need to make sure the search engine parameter is valid and supported by our system.<br><br>"
"I'm preparing the search request with the following configuration:<br>"
"- Search Query: {query}<br>"
"- Search Engine: {engine}<br><br>"
"I'm verifying that the network connection is stable and that the search service is accessible. "
"All preliminary checks have been completed successfully."
),
"executing": (
"I'm now executing the web search for: {query}<br><br>"
"I'm connecting to the {engine} search service and sending the search request. "
"The connection has been established successfully and I'm waiting for the search results. "
"I'm processing multiple search result pages to gather comprehensive information.<br><br>"
"I'm analyzing the search results to identify the most relevant and authoritative sources. "
"The search engine is returning results and I'm filtering them based on relevance scores. "
"I'm extracting key information from each search result including titles, snippets, and URLs.<br><br>"
"I'm organizing the search results in order of relevance and checking for duplicate content. "
"The search process is progressing smoothly and I'm collecting valuable information. "
"I'm also verifying the credibility of the sources to ensure high-quality information.<br><br>"
"Current status: Processing search results...<br>"
"Results found: Multiple relevant sources identified<br>"
"Quality assessment: High relevance detected"
),
"completed": (
"I have successfully completed the web search for: {query}<br><br>"
"I've retrieved comprehensive search results from {engine} and analyzed all the information. "
"The search yielded multiple relevant results that directly address the user's query. "
"I've extracted the most important information and organized it for processing.<br><br>"
"I've identified several high-quality sources with authoritative information. "
"The search results include recent and up-to-date content that is highly relevant. "
"I've filtered out any duplicate or low-quality results to ensure accuracy.<br><br>"
"I'm now processing the collected information to formulate a comprehensive response. "
"The search results provide sufficient detail to answer the user's question thoroughly. "
"I've verified the credibility of the sources and cross-referenced the information.<br><br>"
"Search Summary:<br>"
"- Total results processed: Multiple pages<br>"
"- Relevance score: High<br>"
"- Information quality: Verified and accurate<br>"
"- Sources: Authoritative and recent<br><br>"
"Preview of results:<br>{preview}"
),
"error": (
"I encountered an issue while attempting to search for: {query}<br><br>"
"I tried to execute the web search but encountered an unexpected error. "
"The error occurred during the search process and I need to handle it appropriately. "
"I'm analyzing the error to understand what went wrong and how to proceed.<br><br>"
"Error details: {error}<br><br>"
"I'm attempting to diagnose the issue and considering alternative approaches. "
"The error might be due to network connectivity, service availability, or parameter issues. "
"I will try to recover from this error and provide the best possible response.<br><br>"
"I'm evaluating whether I can retry the search with modified parameters. "
"If the search cannot be completed, I will use my existing knowledge to help the user. "
"I'm committed to providing valuable assistance despite this technical challenge."
)
},
"read_url": {
"parsing": (
"I need to read and extract content from the URL: {url}<br><br>"
"I'm analyzing the URL structure to ensure it's valid and accessible. "
"The URL appears to be properly formatted and I'm preparing to fetch its content. "
"I will extract the main content from this webpage to gather detailed information.<br><br>"
"I'm validating the URL protocol and checking if it uses HTTP or HTTPS. "
"The domain seems legitimate and I'm preparing the request headers. "
"I need to ensure that the website allows automated content extraction.<br><br>"
"I'm configuring the content extraction parameters:<br>"
"- Target URL: {url}<br>"
"- Extraction Method: Full content parsing<br>"
"- Content Type: HTML/Text<br>"
"- Encoding: Auto-detect<br><br>"
"I'm checking if the website requires any special handling or authentication. "
"All preliminary validation checks have been completed successfully."
),
"executing": (
"I'm now accessing the URL: {url}<br><br>"
"I'm establishing a connection to the web server and sending the HTTP request. "
"The connection is being established and I'm waiting for the server response. "
"I'm following any redirects if necessary to reach the final destination.<br><br>"
"I'm downloading the webpage content and checking the response status code. "
"The server is responding and I'm receiving the HTML content. "
"I'm monitoring the download progress and ensuring data integrity.<br><br>"
"I'm parsing the HTML structure to extract the main content. "
"I'm identifying and removing navigation elements, advertisements, and other non-content sections. "
"I'm focusing on extracting the primary article or information content.<br><br>"
"Current status: Extracting content...<br>"
"Response received: Processing HTML<br>"
"Content extraction: In progress"
),
"completed": (
"I have successfully extracted content from: {url}<br><br>"
"I've retrieved the complete webpage content and processed it thoroughly. "
"The extraction was successful and I've obtained the main textual content. "
"I've cleaned the content by removing unnecessary HTML tags and formatting.<br><br>"
"I've identified the main article or information section of the webpage. "
"The content has been properly parsed and structured for analysis. "
"I've preserved important information while filtering out irrelevant elements.<br><br>"
"I'm now analyzing the extracted content to understand its context and relevance. "
"The information appears to be comprehensive and directly related to the topic. "
"I've verified that the content is complete and hasn't been truncated.<br><br>"
"Extraction Summary:<br>"
"- Content length: Substantial<br>"
"- Extraction quality: High<br>"
"- Content type: Article/Information<br>"
"- Processing status: Complete<br><br>"
"Preview of extracted content:<br>{preview}"
),
"error": (
"I encountered an issue while trying to access: {url}<br><br>"
"I attempted to fetch the webpage content but encountered an error. "
"The error prevented me from successfully extracting the information. "
"I'm analyzing the error to understand the cause and find a solution.<br><br>"
"Error details: {error}<br><br>"
"I'm considering possible causes such as network issues, access restrictions, or invalid URLs. "
"The website might be blocking automated access or the URL might be incorrect. "
"I will try to work around this limitation and provide alternative assistance.<br><br>"
"I'm evaluating whether I can access the content through alternative methods. "
"If direct access isn't possible, I'll use my knowledge to help with the query. "
"I remain committed to providing useful information despite this obstacle."
)
}
}
REASONING_DEFAULT = "I'm processing the tool execution request..."
REASONING_DELAY = 0.01 # 10 ms
OS = [
"Windows NT 10.0; Win64; x64",
"Macintosh; Intel Mac OS X 10_15_7",
"X11; Linux x86_64",
"Windows NT 11.0; Win64; x64",
"Macintosh; Intel Mac OS X 11_6_2"
]
OCTETS = [
1, 2, 3, 4, 5, 8, 12, 13, 14, 15,
16, 17, 18, 19, 20, 23, 24, 34, 35, 36,
37, 38, 39, 40, 41, 42, 43, 44, 45, 46,
47, 48, 49, 50, 51, 52, 53, 54, 55, 56,
57, 58, 59, 60, 61, 62, 63, 64, 65, 66,
67, 68, 69, 70, 71, 72, 73, 74, 75, 76,
77, 78, 79, 80, 81, 82, 83, 84, 85, 86,
87, 88, 89, 90, 91, 92, 93, 94, 95, 96,
97, 98, 99, 100, 101, 102, 103, 104, 105, 106,
107, 108, 109, 110, 111, 112, 113, 114, 115, 116,
117, 118, 119, 120, 121, 122, 123, 124, 125, 126,
128, 129, 130, 131, 132, 133, 134, 135, 136, 137,
138, 139, 140, 141, 142, 143, 144, 145, 146, 147,
148, 149, 150, 151, 152, 153, 154, 155, 156, 157,
158, 159, 160, 161, 162, 163, 164, 165, 166, 167,
168, 170, 171, 172, 173, 174, 175, 176, 177, 178,
179, 180, 181, 182, 183, 184, 185, 186, 187, 188,
189, 190, 191, 192, 193, 194, 195, 196, 197, 198,
199, 200, 201, 202, 203, 204, 205, 206, 207, 208,
209, 210, 211, 212, 213, 214, 215, 216, 217, 218,
219, 220, 221, 222, 223
]
BROWSERS = [
"Chrome",
"Firefox",
"Safari",
"Edge",
"Opera"
]
CHROME_VERSIONS = [
"120.0.0.0",
"119.0.0.0",
"118.0.0.0",
"117.0.0.0",
"116.0.0.0"
]
FIREFOX_VERSIONS = [
"121.0",
"120.0",
"119.0",
"118.0",
"117.0"
]
SAFARI_VERSIONS = [
"17.1",
"17.0",
"16.6",
"16.5",
"16.4",
]
EDGE_VERSIONS = [
"120.0.2210.91",
"119.0.2151.97",
"118.0.2088.76",
"117.0.2045.60",
"116.0.1938.81"
]
DOMAINS = [
"google.com",
"bing.com",
"yahoo.com",
"duckduckgo.com",
"baidu.com",
"yandex.com",
"facebook.com",
"twitter.com",
"linkedin.com",
"reddit.com",
"youtube.com",
"wikipedia.org",
"amazon.com",
"github.com",
"stackoverflow.com",
"medium.com",
"quora.com",
"pinterest.com",
"instagram.com",
"tumblr.com"
]
PROTOCOLS = [
"https://",
"https://www."
]
SEARCH_ENGINES = [
"https://www.google.com/search?q=",
"https://www.bing.com/search?q=",
"https://search.yahoo.com/search?p=",
"https://duckduckgo.com/?q=",
"https://www.baidu.com/s?wd=",
"https://yandex.com/search/?text=",
"https://www.google.co.uk/search?q=",
"https://www.google.ca/search?q=",
"https://www.google.com.au/search?q=",
"https://www.google.de/search?q=",
"https://www.google.fr/search?q=",
"https://www.google.co.jp/search?q=",
"https://www.google.com.br/search?q=",
"https://www.google.co.in/search?q=",
"https://www.google.ru/search?q=",
"https://www.google.it/search?q="
]
KEYWORDS = [
"news",
"weather",
"sports",
"technology",
"science",
"health",
"finance",
"entertainment",
"travel",
"food",
"education",
"business",
"politics",
"culture",
"history",
"music",
"movies",
"games",
"books",
"art"
]
COUNTRIES = [
"US", "GB", "CA", "AU", "DE", "FR", "JP", "BR", "IN", "RU",
"IT", "ES", "MX", "NL", "SE", "NO", "DK", "FI", "PL", "TR",
"KR", "SG", "HK", "TW", "TH", "ID", "MY", "PH", "VN", "AR",
"CL", "CO", "PE", "VE", "EG", "ZA", "NG", "KE", "MA", "DZ",
"TN", "IL", "AE", "SA", "QA", "KW", "BH", "OM", "JO", "LB"
]
LANGUAGES = [
"en-US", "en-GB", "en-CA", "en-AU", "de-DE", "fr-FR", "ja-JP",
"pt-BR", "hi-IN", "ru-RU", "it-IT", "es-ES", "es-MX", "nl-NL",
"sv-SE", "no-NO", "da-DK", "fi-FI", "pl-PL", "tr-TR", "ko-KR",
"zh-CN", "zh-TW", "th-TH", "id-ID", "ms-MY", "fil-PH", "vi-VN",
"es-AR", "es-CL", "es-CO", "es-PE", "es-VE", "ar-EG", "en-ZA",
"en-NG", "sw-KE", "ar-MA", "ar-DZ", "ar-TN", "he-IL", "ar-AE",
"ar-SA", "ar-QA", "ar-KW", "ar-BH", "ar-OM", "ar-JO", "ar-LB"
]
TIMEZONES = [
"America/New_York",
"America/Chicago",
"America/Los_Angeles",
"America/Denver",
"Europe/London",
"Europe/Paris",
"Europe/Berlin",
"Europe/Moscow",
"Asia/Tokyo",
"Asia/Shanghai",
"Asia/Hong_Kong",
"Asia/Singapore",
"Asia/Seoul",
"Asia/Mumbai",
"Asia/Dubai",
"Australia/Sydney",
"Australia/Melbourne",
"America/Toronto",
"America/Vancouver",
"America/Mexico_City",
"America/Sao_Paulo",
"America/Buenos_Aires",
"Africa/Cairo",
"Africa/Johannesburg",
"Africa/Lagos",
"Africa/Nairobi",
"Pacific/Auckland",
"Pacific/Honolulu"
]
DESCRIPTION = """
<b>SearchGPT</b> is <b>ChatGPT</b> with real-time web search capabilities and the ability to read content directly from a URL.
<br><br>
This Space implements an agent-based system with <b><a href="https://www.gradio.app" target="_blank">Gradio</a></b>. It is integrated with
<b><a href="https://docs.searxng.org" target="_blank">SearXNG</a></b>, which is then converted into a script tool or function for native execution.
<br><br>
The agent mode is inspired by the <b><a href="https://openwebui.com/t/hadad/deep_research" target="_blank">Deep Research</a></b> from
<b><a href="https://docs.openwebui.com" target="_blank">OpenWebUI</a></b> tools script.
<br><br>
The <b>Deep Research</b> feature is also available on the primary Spaces of <b><a href="https://umint-openwebui.hf.space"
target="_blank">UltimaX Intelligence</a></b>.
<br><br>
Please consider reading the <b><a href="https://huggingface.co/spaces/umint/ai/discussions/37#68b55209c51ca52ed299db4c"
target="_blank">Terms of Use and Consequences of Violation</a></b> if you wish to proceed to the main Spaces.
<br><br>
<b>Like this project? Feel free to buy me a <a href="https://ko-fi.com/hadad" target="_blank">coffee</a></b>.
""" # Gradio