GPT-Crawler builds GPTs knowledge base with one-click crawler

write at the front

GPT-Crawler builds GPTs knowledge base with one-click crawler
It can crawl website data and build GPTs knowledge base. The project depends on the node.js environment. Next, we follow the steps to install it. , very simple

Reference: https://zhuanlan.zhihu.com/p/668700619


In the era of information explosion, data has become the new oil. However, how to effectively extract valuable knowledge from this endless network information has become a major challenge faced by technical personnel. Especially for advanced technologies like GPTs, building a strong and updated knowledge base is crucial. This is the mission of the GPT-Crawler one-click crawler tool we are going to introduce today - a powerful tool designed to help developers and data scientists efficiently build and maintain the knowledge base of GPTs.

In this blog, we will delve into how to use GPT-Crawler to capture and process network data to provide rich and accurate information for GPTs models. Starting from installing Node.js as the basis of the running environment, we will guide you step by step on how to install and start GPT-Crawler. Additionally, we’ll cover how to combine this powerful crawler tool with OpenAI’s technology and how to customize assistants and GPTs to meet your specific needs and preferences.

Whether you are an experienced developer or a beginner with a strong interest in data science and artificial intelligence, this blog will provide you with valuable knowledge and practical guidance. By reading this article, you will not only learn how to build an efficient GPTs knowledge base, but also gain practical tips on how to customize and optimize crawlers. So, let’s start this exciting technology journey together!

Install node.js

Node.js download address: https://nodejs.org/en
Just download version 20.10.0, and it will be installed by default after downloading

Insert image description here

Enter node -v on the command line after the installation is complete. If the version is displayed, the installation is successful

Insert image description here

Install GPT-Crawler

Project address: https://github.com/BuilderIO/gpt-crawler
This project can crawl website data and generate knowledge base files for creating GPTs
After opening the project address, click [Code] to download the compressed file, save it to your computer and decompress it locally

VSCode official website download: https://code.visualstudio.com/Download

Open with VSCode encoding tool,
or click on VSCode, file in the upper left corner, click to open the folder, and select the decompression address

Switch to the project directory (for example, mine is C:\Users\Yu\Desktop\gpt-crawler-main\gpt-crawler-main>), right-click and select Open in integrated terminal< a i=1> Enter to install the project dependency package
npm install

Insert image description here

Start the crawler

npm start

The following log output is normal:

Insert image description here

Finally, the program will output the file output.json in the project root directory, which is the file we need.
The output file is in this directory
Click it to take a look, it’s perfect

Insert image description here

Combined with OpenAI

This step requires an OpenAI account, and we need to upload the generated file to OpenAI.

Custom assistant

The advantage of choosing assistant is that we can use the assistant API provided by OpenAI to integrate it into our own system.

Operation steps:
1. Enter the custom Assistants page https://platform.openai.com/assistants
2. Create an Assistant< /span> 4. Configure other options
3. Add the output.json file generated above

After uploading the configuration, click Save and then start testing:

Custom GPTs (the one I use)

Customizing GPTs is similar to the above operation, so you can experience it yourself. GPTs requires everyone to pay Plus, and it seems that the official API for GPTs is not yet available.

Summarize

The GPT Crawler project allows us to customize our own knowledge base with only simple configuration. Combined with OpenAI's API, a lot of things can be done. Everyone can experience it by themselves.

Guess you like

Origin blog.csdn.net/wtyuong/article/details/134819013