OpenAI's New Built-in Tools: Towards Agentic Systems
OpenAI has recently introduced a suite of built-in tools that extend the capabilities of its AI models, making it easier for developers to build powerful AI-driven applications. These tools are part of OpenAI’s evolving platform (notably the new Responses API and Agents SDK) geared towards creating agentic AI systems – AI agents that can independently perform tasks on behalf of users. In this article, we provide an overview of these new tools, explain their purpose and how they enhance AI applications, explore real-world use cases, and demonstrate with code examples how to use them. We’ll also cover best practices to implement these tools effectively in your own projects.
Overview of OpenAI's Built-in Tools and Their Purpose
OpenAI’s built-in tools allow AI models to access external information or perform actions, overcoming some limitations of standalone LLMs. The current set of new tools includes:
Each tool serves a specific purpose: Web Search brings in fresh knowledge and transparency (via citations), File Search injects domain-specific or proprietary data into the AI’s context, and Computer Use grants the AI the ability to take actions in digital environments. Together, these tools greatly enhance what AI applications can do, enabling more useful and reliable agents that can retrieve information and take actions autonomously.
Web Search Tool – Real-time Information with Citations
The Web Search tool gives AI models access to the vast information on the internet. Instead of relying solely on the model’s pre-trained knowledge, the AI can perform live web queries and incorporate the results into its responses. This means the model can fetch fast, up-to-date answers from the web, complete with relevant source links. The answers generated through this tool include inline citations (e.g. links to news articles or websites), allowing users to verify facts and explore sources further. This transparency builds trust and enables new user experiences where the AI not only provides information but also shows where that information came from.
How it enhances applications: With web search, AI applications are no longer limited by a static knowledge cutoff. They can handle queries about recent news, live sports scores, stock prices, or any timely data. For example, an AI assistant could answer “What happened in tech news this morning?” by searching the web and giving an answer with references to news sites. This opens the door to applications like real-time research assistants, shopping advisors, or travel planners that always use the latest information. During early testing, developers have built shopping assistants, research agents, and travel booking agents – essentially any use-case requiring timely web info can benefit from this tool.
Real-world example: One company leveraging web search is Hebbia. Hebbia integrates OpenAI’s web search tool to help financial and legal professionals quickly extract insights from extensive public and private data sources. By incorporating real-time web search into their workflow, Hebbia’s AI agent can deliver richer, context-specific market intelligence and consistently improve the relevance of its analyses, outperforming previous benchmarks. In essence, web search empowers such agents to be significantly more knowledgeable and up-to-date.
Using the Web Search tool – step-by-step example: OpenAI’s API provides an easy way to use the web search tool in your requests. Under the hood, this tool is available when using the gpt-4o or gpt-4o-mini model in the new Responses API (these are fine-tuned models optimized for tool use). Here’s how you can use the web search tool in practice:
For example, using the OpenAI Python library (or an equivalent HTTP request), a call might look like this:
import openai
openai.api_key = "YOUR_API_KEY"
# Ask a question with the web search tool enabled
response = openai.Responses.create(
model="gpt-4o",
tools=[{"type": "web_search_preview"}],
input="What was a positive news story that happened today?"
)
print(response.output_text)
In this snippet, we request GPT-4 to find a positive news story from today. The model will perform a web search under the hood and then produce an answer. The printed output_text might be a summary of the news story along with a citation link to the source. For example, it could return something like:
"One positive news story today is that scientists announced a breakthrough in renewable energy storage, which could accelerate the adoption of clean energy."
The citation in the text (shown here as a reference) would point to a specific article or source supporting the statement. By including citations, the web search tool not only gives the latest information but also enhances trust and verifiability in AI-generated content.
File Search Tool – Leveraging Custom Data and Documents
The File Search tool allows your AI assistant to ingest and retrieve information from your own documents or knowledge base. This is extremely useful for enterprise and domain-specific applications, where answers need to be based on proprietary data (such as product documentation, internal wikis, PDFs of research papers, etc.). OpenAI’s file search works by indexing documents in a vector store and then querying that store for relevant content when the model receives a prompt. Under the hood, it uses embeddings and similarity search to find text segments that are relevant to the query, and makes those available to the model.
How it enhances applications: By using file search, developers can build retrieval-augmented generation (RAG) pipelines with very little effort. The model can pull in knowledge from large volumes of text that it otherwise wouldn’t know. This means an AI app can have up-to-date company policies, product specs, manuals, or any corpus of choice at its fingertips. The file search tool supports multiple file types and offers features like query optimization, metadata filtering, and result reranking for accuracy (New tools for building agents | OpenAI). Essentially, it brings enterprise knowledge integration to AI: a customer support bot can fetch answers from FAQ documents, a legal assistant bot can cite relevant case files, or a coding helper can look up an API reference – all through the same unified model interface.
Use cases and benefits: There are many real-world scenarios for file search. For instance, a customer support agent AI could instantly retrieve answers from a company’s FAQ or knowledge base, providing customers with precise information. A legal research assistant AI might quickly scan a database of court cases or regulations to find relevant passages for a lawyer’s query. A developer’s coding assistant could search through documentation or a codebase to help answer technical questions. In fact, the tool has been used to enable things like a travel agent AI retrieving company travel policy details for users. One example is Navan, which uses the file search tool in its AI travel assistant to provide users with precise answers from internal knowledge-base articles, such as a company’s travel policy. Thanks to built-in query tuning and reranking, Navan set up a powerful RAG system without extra custom code or tuning. By maintaining dedicated vector stores for different user groups, they even tailor answers to each customer’s context, resulting in more accurate and personalized support. This demonstrates how file search can save time and improve accuracy for end-users by drawing on the right piece of information at the right moment.
Using the File Search tool – step-by-step example: To use file search, you need to first prepare your document data and then make a query with the tool. Here’s a typical workflow:
For example, here’s how you might implement a file search query in Python:
Recommended by LinkedIn
# (Assume openai.api_key is already set and you have file IDs from previously uploaded files)
# 1. Create a vector store for your documents
product_docs = openai.VectorStore.create(
name="Product Documentation",
file_ids=[file1_id, file2_id, file3_id] # IDs of documents to index
)
# 2. Use the file_search tool in a model request
response = openai.Responses.create(
model="gpt-4o-mini",
tools=[{
"type": "file_search",
"vector_store_ids": [product_docs.id] # reference the created vector store by ID
}],
input="What is 'Deep Research' in the context of OpenAI?" # sample user query
)
# 3. Output the result
print(response.output_text)
In this code, we first create a vector store named "Product Documentation" containing three files (perhaps PDFs or text files of product research). Then we ask the model a question: “What is 'Deep Research' by OpenAI?” including the file search tool with our product_docs store. The model will search those documents for the term "Deep Research" and related context. The output_text we get back might be something like:
“‘Deep Research’ is an OpenAI initiative aimed at long-term fundamental research in AI, focusing on high-risk, high-reward ideas. According to OpenAI’s documentation, Deep Research involves collaborations with academic institutions and exploration of new learning paradigms.”
If that information was contained in one of the provided documents, the model can extract and present it. In practice, the answer might also cite the document title or source. By using file search, we ensure the assistant’s answer is grounded in the actual data we provided, which greatly improves accuracy and usefulness in domain-specific applications.
Computer Use Tool – Automating Actions in an Environment
The Computer Use tool is perhaps the most groundbreaking of the new built-in tools. It gives the AI the ability to perform actions on a computer — essentially controlling a virtual computer interface (like moving a mouse, clicking, typing, scrolling). This is powered by OpenAI’s Computer-Using Agent (CUA) model, a specialized model that can interpret instructions and carry out sequences of GUI operations. In the OpenAI API, the computer use tool captures the model’s suggested mouse and keyboard actions and translates them into executable commands in a controlled environment (New tools for building agents | OpenAI). In simpler terms, the AI can operate software or websites for you, under supervision.
How it enhances applications: This capability turns an AI model into an agent that can not only think or generate text, but also act. It can handle tasks that typically require a human using a computer. For example, an AI agent could fill out forms on a website, navigate through a web app, or cross-reference information by clicking through pages. This is extremely useful for automating workflows where no convenient API exists. Developers can use it to perform browser-based tasks like web app testing, data entry in legacy systems, or any repetitive online operation. Essentially, it’s like an AI-powered Robotic Process Automation (RPA) but driven by natural language instructions.
Use cases and real-world examples: The computer use tool can be applied to a wide range of scenarios:
By enabling these use cases, the computer use tool effectively allows AI to bridge the gap between language understanding and real-world action. It enhances AI applications by letting them interact with the same interfaces humans use, which is a significant step toward more autonomous agents.
Using the Computer Use tool – step-by-step example: The computer use tool is currently available as a research preview (to select developers in higher usage tiers), which means you may need special access to experiment with it. Assuming you have access, here’s how you can use it:
Here is an illustrative example using the OpenAI API (pseudo-code in Python format):
# Using the computer use tool to have the AI perform a task in a browser.
response = openai.Responses.create(
model="computer-use-preview",
tools=[{
"type": "computer_use_preview",
"display_width": 1024,
"display_height": 768,
"environment": "browser"
}],
input="I'm looking for a new camera. Help me find the best one."
)
print(response.output)
In this example, the AI is instructed to help find the best camera. What happens behind the scenes is the model will likely open a browser (in the sandboxed environment), perform a search for cameras, navigate through search results or shopping sites, and compile the information. The response.output could be a structured log or description of what it did (for instance, “Opened example.com, searched for ‘best camera 2025’, clicked the first result, price noted was $X”). It might also return a synthesized answer like, “The XYZ Camera is highly rated this year for its image quality and is available for $799 on CameraShop.com.” along with details of how it found that info.
Safety and control: Because this tool actually carries out actions, OpenAI has built in multiple safeguards. The model’s actions are executed in a controlled environment (not on your actual machine). OpenAI’s research showed the CUA model achieves impressive results on benchmarks for computer tasks (e.g., setting a new state-of-the-art on the WebArena browsing benchmark), but it’s not perfect and can make mistakes. Thus, the platform has safety checks and confirmations in place. For instance, the system may require the AI to get user confirmation before performing sensitive operations (like deleting data or making a purchase), and there are filters to guard against malicious prompt injections that might try to exploit the tool (Openai Agents SDK, Responses Api Tutorial - DEV Community). As a developer, you should also implement checks on your side – for example, reviewing the actions an agent wants to take, and sandboxing the environment (which OpenAI enables) so no unintended harm can occur. We’ll discuss more best practices on this below.
Best Practices for Implementing Built-in Tools Effectively
Using OpenAI’s built-in tools can greatly enhance your AI application, but to get the best results and ensure safety, keep these best practices in mind:
By following these practices, you can harness OpenAI’s built-in tools to build more powerful, safe, and reliable AI applications. Whether you’re creating an AI assistant that surfs the web for answers, a chatbot that consults enterprise knowledge bases, or an autonomous agent that navigates software interfaces, these tools provide the building blocks to do so with minimal overhead.
Conclusion
OpenAI's new built-in tools mark an exciting advancement in the AI developer toolkit. They effectively extend an AI model’s abilities beyond text generation to include searching knowledge (Web Search), leveraging custom data (File Search), and taking actions (Computer Use). This built-in functionality can dramatically improve what AI applications can do – enabling them to provide up-to-date information, grounded answers from proprietary data, and even automate tasks in the digital world. The introduction of these tools (alongside the Responses API and Agents SDK) has streamlined the core logic and orchestration required to build AI agents, making it significantly easier for developers to get started.
As we have seen, the potential use cases are vast: from enhancing customer support and personal assistants to automating business processes and beyond. By incorporating the best practices outlined above – choosing the right tool, guiding the model properly, and implementing safety checks – developers and AI enthusiasts can build innovative applications that were not previously possible with standalone GPT-style models. OpenAI continues to improve these tools (currently some are in preview) and plans to introduce even more capabilities in the future..
In summary, OpenAI’s built-in tools empower you to build AI systems that learn, look up information, and act. With careful implementation, they can unlock a new generation of AI applications that are more knowledgeable, context-aware, and action-oriented. Whether you're building a smart research agent or an autonomous helper for tedious tasks, these tools provide a powerful foundation to bring your AI ideas to life. Happy building!