chore: docs

This commit is contained in:
Exceeds_Lenovo\rohan 2025-04-23 15:22:43 -07:00
parent 7044eae027
commit f445a2b130
5 changed files with 29 additions and 17 deletions

View file

@ -45,7 +45,23 @@ nmake /F Makefile.win install
See the [installation notes](https://github.com/pgvector/pgvector/tree/master#installation-notes---windows) if you run into issues
---
## Google OAuth Setup
SurfSense user management and authentication works on Google OAuth. Lets set it up.
1. Login to your [Google Developer Console](https://console.cloud.google.com/)
2. Enable People API.
![Google Developer Console People API](/docs/google_oauth_people_api.png)
3. Set up OAuth consent screen.
![Google Developer Console OAuth consent screen](/docs/google_oauth_screen.png)
4. Create OAuth client ID and secret.
![Google Developer Console OAuth client ID](/docs/google_oauth_client.png)
5. It should look like this.
![Google Developer Console Config](/docs/google_oauth_config.png)
---
## File Upload's
@ -53,34 +69,30 @@ Files are converted to LLM friendly formats using [Unstructured](https://github.
1. Get an Unstructured.io API key from [Unstructured Platform](https://platform.unstructured.io/)
2. You should be able to generate API keys once registered
![Image](/docs/unstructured.png)
![Unstructured Dashboard](/docs/unstructured.png)
## Google OAuth Setup
SurfSense user management and authentication works on Google OAuth.
1. Set up OAuth Client at [Google Developer Console](https://developers.google.com/identity/protocols/oauth2)
2. Enable People API.
3. Note your OAuth Client.
---
## LLM Observability (Optional)
For monitoring LLM interactions:
This is not required for SurfSense to work. But it is always a good idea to monitor LLM interactions. So we do not have those WTH moments.
1. Get a LangSmith API key from [smith.langchain.com](https://smith.langchain.com/)
2. This helps in observing SurfSense Researcher Agent operations
2. This helps in observing SurfSense Researcher Agent.
![LangSmith](/docs/langsmith.png)
## Crawler Support
---
SurfSense currently uses [Firecrawl.py](https://www.firecrawl.dev/) for web crawling.
## Crawler
## API Keys
SurfSense have 2 options for saving webpages:
- [SurfSense Extension](https://github.com/MODSetter/SurfSense/tree/main/surfsense_browser_extension) (Overall better experience & ability to save private webpages, recommended)
- Crawler (If you want to save public webpages)
Prepare the following API keys:
**NOTE:** SurfSense currently uses [Firecrawl.py](https://www.firecrawl.dev/) for web crawling. If you plan on using the crawler, you will need to create a Firecrawl account and get an API key.
1. LLM API keys (OpenAI, Gemini, etc.) based on your selected models
2. Unstructured.io API key
3. Firecrawl API key (if using the crawler)
---
## Next Steps

Binary file not shown.

After

Width:  |  Height:  |  Size: 47 KiB

View file

Before

Width:  |  Height:  |  Size: 68 KiB

After

Width:  |  Height:  |  Size: 68 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 71 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 301 KiB