Deep Research Pipeline with Google Drive Connector
Overview
This section demonstrates Pipeline orchestration with additional data sources, using Google Drive as an example.
In this setup:
the Pipeline controls routing and execution flow
Google Drive access is provided via an MCP connector
Deep research is invoked as an encapsulated step with additional data available during execution
You can either:
You can refer to the guide whenever you need explanation or want to clarify how each part works.
Follow along with each step to recreate the files yourself while learning about the components and how to integrate them.
Both options will work—choose based on whether you prefer speed or learning by doing!
Prerequisites
This example specifically requires:
Completion of all setup steps listed on the Prerequisites page.
You should be familiar with these:
Event emitting: Event Emitter
Installation
Project Setup
Clone the repository
Set UV authentication and install dependencies
Unix-based systems (Linux, macOS):
For Windows:
Prepare .env file with Google Drive authentication
Get the auth token from OpenAI Connector Guide
When generating the token, make sure to enable the following scopes:
userinfo.emailuserinfo.profiledrive.readonly
Add to .env:
Implementation
In this example, Google Drive is made available to deep research through an MCP connector, while the Pipeline determines when deep research should be invoked.
The Google Drive connector extends the data available during research, but does not change the execution flow or reasoning strategy of deep research itself.
Run the script
Benefits:
Make documents stored in Google Drive available as research context
Combine private documents with public information during research
Integrate external data sources without changing research execution logic
Next Step
Combine in an end-to-end RAG pipeline by following the guides in Your First RAG Pipeline
Explore more about deep researcher subclasses and features in API reference page
Last updated
Was this helpful?