Save as New
Run
Examples
API
Integrations
GPT-4o (openai)
Submit Links in Bulk
⚙️ Settings
How should the LLM interpret the results from your knowledge base?
Numbers ( [1] [2] [3] ..)
🔗 Shorten Citation URLs
To improve answer quality, pick a synthetic data maker workflow to scan & OCR any images in your documents or transcribe & translate any videos. It also can synthesize a helpful FAQ. Adds ~2 minutes of one-time processing per file.
———
In general, you should not need to adjust these.
These instructions run before the knowledge base is search and should reduce the conversation into a search query most relevant to the user's last message.
Instructions to create a query for keyword/hybrid BM25 search. Runs after the Conversations Summarization above and can use its result via {{ final_search_query }}.
Text Embedding 3 Large (OpenAI)
Weightage for dense vs sparse embeddings. 0 for sparse, 1 for dense, 0.5 for equal weight.Generally speaking, dense embeddings excel at understanding the context of the query, whereas sparse vectors excel at keyword matches.
0
1
0.5
The maximum number of document search citations.
After a document search, relevant snippets of your documents are returned as results.This setting adjusts the maximum number of words in each snippet (tokens = words * 2).A high snippet size allows the LLM to access more information from your document results, at the cost of being verbose and potentially exhausting input tokens (which can cause a failure of the copilot to respond).
Your knowledge base documents are split into overlapping snippets.This settings adjusts how much those snippets overlap (overlap tokens = snippet tokens / overlap ratio).In general you shouldn't need to adjust this.
Avoid Repetition
The maximum number of tokens to generate in the completion. Increase to generate longer responses.
Higher values allow the LLM to take more risks. Try values larger than 1 for more creative applications or 0 to ensure that LLM gives the same answer when given the same user input.
How many answers should the copilot generate? Additional answer outputs increase the cost of each run.
Run cost = 13 credits
Breakdown: 10 (GPT-4o (openai)) + 3/run
By submitting, you agree to Gooey.AI's terms & privacy policy.
Show Raw Output
Assistant
Determining the "best" vector database can be subjective and depends on specific use cases, requirements, and preferences. However, based on the discussions from the search results, here are some insights into various vector databases and their strengths:
Pinecone:
Weaviate:
Qdrant:
Chroma:
Redis:
Pgvector (Postgres with vector extension):
In conclusion, the best vector database depends on your specific needs:
UserWhat is the best vector db?
📎
✈ Send
🗑️ Clear
💁♀️ Sources
Regenerate
Generated in 14.9s on
...
ℹ️ Details
🙋🏽♀️ Need more help? Join our Discord
Create realistic lipsync videos with custom voices. Just upload a video or image, choose a voice from Google, OpenAI or bring your own voice from Eleven Labs to generate amazing videos with the Gooey.AI …
We've built the best Retrieval Augmented Generation (RAG) as-a-Service anywhere - now with page-level citations! Absorb tables, PDFs, docs, links, videos or audio clips and use our synthetic data maker to …
Create AI-generated Animation without relying on complex CoLab notebooks. Input your prompts + keyframes and bring your ideas to life using the animation capabilities of Gooey & Stable Diffusion's Deforum. …
Create multiple AI photos from one prompt using Stable Diffusion (1.5 -> 2.1, Open/Midjourney), DallE, and other models. Find out which AI Image generator works best for your text prompt on comparing OpenAI, …