Build your AI apps & agents faster
Build your AI apps & agents faster
Batteries-included,
Serverless RAG-as-a-Service Platform
Supports all OpenAI, Anthropic, Cohere, Google AI, Azure AI,
Mistral, Deepseek, Jina, Voyage, Groq and Cerebras LLMs and embedding models
Automated data preparation for multimodal content, including Markdown extraction, audio transcription and image analysis
Built-in data connectors for web scraping, Google Drive, Notion, GitHub, Slack, Jira, email, RSS and many more
Batteries-included,
Serverless RAG-as-a-Service Platform
Supports all OpenAI, Anthropic, Cohere, Google AI, Azure AI, Mistral,
Deepseek, Jina, Voyage, Groq and Cerebras LLMs and embedding models
Automated data preparation for multimodal content, including Markdown extraction, audio transcription and image analysis
Built-in data connectors for web scraping, Google Drive,
Notion, GitHub, Slack, Jira, email, RSS and many more
Batteries-included,
RAG-as-a-Service Platform
Supports all OpenAI, Anthropic, Cohere, Google AI, Azure AI, Mistral, Deepseek, Jina, Voyage, Groq and Cerebras LLMs and embedding models
Automated data preparation for multimodal content, including Markdown extraction, audio transcription and image analysis
Built-in data connectors for web scraping, Google Drive, Notion, GitHub, Slack, Jira, email, RSS and many more
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
await graphlit.client.ingest_uri(
uri="https://www.graphlit.com"
)
response = await graphlit.client.prompt_conversation(
prompt="How can Graphlit accelerate my Generative AI app development?"
)
message = response.prompt_conversation.message.message
print(message)
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
await graphlit.client.ingest_uri(
uri="https://www.graphlit.com"
)
response = await graphlit.client.prompt_conversation(
prompt="How can Graphlit accelerate my GenAI app development?"
)
message = response.prompt_conversation.message.message
print(message)
Start building today with Next.js and Graphlit!
Start building today with Next.js and Graphlit!
Chat with your Documents
Extract Markdown from Documents
Scrape Websites
What can I ingest?
Ingest from any data source such as:
Web sites, cloud storage, SharePoint, podcasts,
Jira, Notion, YouTube, email or Slack
Ingest any unstructured data format such as:
documents, HTML, Markdown, audio, video, or images
AUTOMATED ETL FOR LLMs
AUTOMATED ETL FOR LLMs
AUTOMATED ETL FOR LLMs
High-performance
data preparation
High-performance
data preparation
High-performance
data preparation
Data feeds for automated ingestion
Extract text and tables from documents and images with OCR and LLMs
Automatic audio transcription with Deepgram
Automated web scraping
Enrich data with external APIs, such as Wikipedia and Crunchbase
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
input = FeedInput(
name=f"{account-name}: {container-name}",
type=FeedTypes.SITE,
site=SiteFeedPropertiesInput(
type=FeedServiceTypes.AZURE_BLOB,
isRecursive=True,
azureBlob=AzureBlobFeedPropertiesInput(
accountName="{account-name}",
containerName="{container-name}",
storageAccessKey="{storage-key}",
prefix="{prefix}"
)
),
workflow=EntityReferenceInput(
id="{workflow-id}"
)
)
response = await graphlit.client.create_feed(input)
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
input = FeedInput(
name=f"{account-name}: {container-name}",
type=FeedTypes.SITE,
site=SiteFeedPropertiesInput(
type=FeedServiceTypes.AZURE_BLOB,
isRecursive=True,
azureBlob=AzureBlobFeedPropertiesInput(
accountName="{account-name}",
containerName="{container-name}",
storageAccessKey="{storage-key}",
prefix="{prefix}"
)
),
workflow=EntityReferenceInput(
id="{workflow-id}"
)
)
response = await graphlit.client.create_feed(input)
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
input = WorkflowInput(
name="Azure AI Document Intelligence",
preparation=PreparationWorkflowStageInput(
jobs=[
PreparationWorkflowJobInput(
connector=FilePreparationConnectorInput(
type=FilePreparationServiceTypes.AZURE_DOCUMENT_INTELLIGENCE,
azureDocument=AzureDocumentPreparationPropertiesInput(
model=AzureDocumentIntelligenceModels.LAYOUT
)
)
)
]
)
)
response = await graphlit.client.create_workflow(input)
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
input = WorkflowInput(
name="Azure AI Document Intelligence",
preparation=PreparationWorkflowStageInput(
jobs=[
PreparationWorkflowJobInput(
connector=FilePreparationConnectorInput(
type=FilePreparationServiceTypes.AZURE_DOCUMENT_INTELLIGENCE,
azureDocument=AzureDocumentPreparationPropertiesInput(
model=AzureDocumentIntelligenceModels.LAYOUT
)
)
)
]
)
)
response = await graphlit.client.create_workflow(input)
RAG-AS-A-Service
RAG-AS-A-Service
RAG-AS-A-Service
What else?
What else?
What else?
RAG and GraphRAG Ready: Intelligent text extraction and chunking, built-in vector embeddings and conversation history, LLM-based entity extraction
Semantic Search: Vector-based search, including metadata filtering
Content Creation: Automated text and transcript summarization, social media post generation, long-form content creation
ANY CONTENT, ANY FORMAT
ANY CONTENT, ANY FORMAT
ANY CONTENT, ANY FORMAT
Multimodal ready
Multimodal ready
Multimodal ready
Integrated with Large Multimodal Models (LMMs) including OpenAI
GPT-4o and Anthropic Sonnet 3.5Generate image descriptions with visual object detection
Similarity search via image embeddings
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
response = await graphlit.client.query_contents(
filter=ContentFilter(
search_type=VECTOR,
search="Unstructured data",
types=[ FILE, PAGE ],
contents=[
EntityReferenceFilter(
id="{content-id}"
)
]
)
)
from graphlit import Graphlit
from graphlit_api import *
graphlit = Graphlit()
input = WorkflowInput(
name="Azure AI Document Intelligence",
preparation=PreparationWorkflowStageInput(
jobs=[
PreparationWorkflowJobInput(
connector=FilePreparationConnectorInput(
type=FilePreparationServiceTypes.AZURE_DOCUMENT_INTELLIGENCE,
azureDocument=AzureDocumentPreparationPropertiesInput(
model=AzureDocumentIntelligenceModels.LAYOUT
)
)
)
]
)
)
response = await graphlit.client.create_workflow(input)
BUILT FOR DEVELOPERS, BY DEVELOPERS
BUILT FOR DEVELOPERS, BY DEVELOPERS
BUILT FOR DEVELOPERS, BY DEVELOPERS
Easy integration
Easy integration
Easy integration
Native SDKs for Python, Node.js, .NET
No infrastructure to be deployed
Integrated usage logs
Serverless, cloud-native platform
Multitenant-ready with RBAC
Data is encrypted-at-rest
Usage-based pricing
Our platform.
Your apps & agents.
Any unstructured data.
Our platform.
Your apps & agents.
Any unstructured data.
For all developers building AI apps & agents
For developers building chatbots, copilots, or vertical AI applications
with domain-specific data
For all developers building AI apps & agents
Pricing
How much does Graphlit cost?
Free to get started, no credit card required.
With our paid tiers, your costs are based on how much content you ingest -
plus any LLM token usage cost, such as RAG conversations.
Usage-based pricing starts at $0.10/credit
Let's chat!
Since every application is different,
we are happy to model out costs for your use case.
Free to get started, no credit card required.
With our paid tiers, your costs are based on how much content you ingest - plus any LLM token usage cost, such as RAG conversations.
Usage-based pricing starts at $0.10/credit
Let's chat!
Since every application is different, we are happy to model out costs for your use case.
Content cost
PDF: 5 pages/credit
PDF: 5 pages/credit
Audio: 10 minutes/credit
Audio: 10 minutes/credit
DOCX: 5 pages/credit
DOCX: 5 pages/credit
Video: 4 minutes/credit
Video: 4 minutes/credit
PPTX: 5 slides/credit
PPTX: 5 slides/credit
Images: 15 images/credit
Images: 15 images/credit
Web pages: 5 pages/credit
Web pages: 5 pages/credit
Markdown, Plain Text: 3 credits/Mtok
Markdown: 3 credits/Mtok
Markdown: 3 credits/Mtok
Costs based on averages. Your cost may vary depending on the specific content.
Costs based on averages. Your cost may vary depending on the specific content.
Costs based on averages.
Your cost may vary depending on the specific content.
Free
$0
per/month
Ingest any content type
(i.e. PDFs, MP3s, web pages)
Create content feeds
(i.e. RSS, Web, Notion, blob storage)
Search content by
text or vector similarity
Filter content by metadata
Create chatbot conversations
over your content
Configure content workflows
Includes Deepgram audio transcription
Includes all vector embeddings
and prompt completions
Supports multi-tenant apps
Includes 100 credits
Includes 1GB content storage
Includes 1000 content items
Includes 3 feeds
Includes 100 chatbot
conversations
Community Discord support
Hobby
$49
/month + usage
Everything in Free tier
$0.10/credit usage
Includes 10GB content storage
Includes 10K content items
Unlimited feeds
Unlimited chatbot conversations
Email and community Discord support
Starter
$199
/month + usage
Everything in Hobby tier
$0.09/credit usage (10% off)
Includes 100GB content storage
Includes 100K content items
Unlimited feeds
Unlimited chatbot conversations
Priority email, private Slack support
Growth
$999
/month + usage
Everything in Starter tier
$0.08/credit usage (20% off)
Unlimited content storage
Unlimited content items
Unlimited feeds
Unlimited chatbot conversations
Priority email, private Slack support
Dedicated technical contact
SLA (coming soon)
SOC 2 (coming soon)
Definitions:
Content: Any ingested file, web page, Slack message, email, etc.
Feed: Any automated data ingestion from Web site, SharePoint, S3 bucket, etc.
Conversation: Any threaded conversation with LLM
Credit: Aggregated unit of serverless cloud compute, cloud storage,
LLM tokens and third-party API usage
Definitions:
Content: Any ingested file, web page, Slack message, email, etc.
Feed: Any automated data ingestion from Web site, SharePoint, S3 bucket, etc.
Conversation: Any threaded conversation with LLM
Credit: Aggregated unit of serverless cloud compute, cloud storage,
LLM tokens and third-party API usage
Newsletter
Get Graphlit updates to your inbox
Contact Us
Got questions?
We would be happy to talk about what you're building.
Email [email protected], or schedule a call below.
We would be happy to talk about what you're building.
Email [email protected], or
join us on Discord and let's chat.
We would be happy to talk about what you're building.
Email [email protected], or
join us on Discord and let's chat.
Can I use this to build a chatbot or copilot?
Definitely! Graphlit provides everything you need to build a RAG application, such as a chatbot or copilot, with our easy-to-use API. Compared to the OpenAI Assistants API, Graphlit handles a wider range of content formats, and offers higher limits on storage capacity. Also, we provide vector-based semantic search as well as RAG conversations. In addition to a small monthly platform fee, you are charged for credit usage based on the volume of content ingested into the platform, LLM tokens used, and other cloud API usage. But you can also bring your own LLM keys, and pay for those tokens yourself.
Can I extract text from PDFs or Word documents?
Yes! You can ingest any type of document, including PDF, Word document, Powerpoint presentation, Markdown, etc. and Graphlit will extract structured text. You can optionally use a content workflow with Azure AI Document Intelligence models, for high-quality OCR text and table extraction. In our testing, we have seen comparable or better results using Azure AI Document Intelligence for OCR, compared to Unstructured.IO, LlamaParse, and other open source PDF extractors.
How is this different than LangChain?
LangChain is a leading frameworks for building LLM applications, but it is not a managed platform. When building RAG applications, you would need to integrate LangChain with LLMs, vector databases, and cloud infrastructure. There are many pieces to assemble to have a production-grade application with LangChain, and you are left to solve these DevOps problems yourself. WIth Graphlit, you can start building your RAG application immediately - no assembly required. We provide a scalable, managed API, where you just have to point us at your unstructured data, and we do the rest. We handle text extraction, text chunking, vector embeddings, as well as RAG conversations (with history). Also, Graphlit handles multi-tenant semantic search out of the box.
How is this different than Unstructured.IO?
Unstructured.IO is focused on the extraction of text from PDFs and other documents, and the partitioning of those documents. They are not a RAG-ready platform, in that they don't directly connect your data to LLMs. They are also not multimodal-ready. They are currently focused on textual formats, not any other media formats. You still need to build an end-to-end unstructured data pipeline around their technology. By using Graphlit, you can start building multimodal RAG applications immediately - no assembly required. We offer best-in-class OCR and text extraction, using Azure AI Document Intelligence. Graphlit also supports a wider range of file formats and metadata than Unstructured.IO, and has built-in audio transcription.
Do I need my own vector database, like Pinecone or Qdrant?
Nope! With Graphlit, you get everything you need to create RAG applications out of the box. We already have integrated best-in-class vector search and vector embeddings into the platform.
Do you support a JSON mode for text extraction?
Yes! Graphlit stores extracted text and tables in our own JSON format, which you can access through our API. You have the option of using Graphlit for ingestion and text extraction only, and then post-processing the JSON format with your own application. But in most cases, you would have Graphlit generate vector embeddings from the extracted text, and handle the RAG conversations for you.
Do you support audio and video transcription?
For sure, we handle a variety of audio and video formats, such as MP4, MP3, WAV and AAC/M4A. Any audio or video file ingested into Graphlit will have audio automatically transcribed using the latest Deepgram speech-to-text model.
Can I build a knowledge graph from my data?
Certainly! Graphlit automatically builds a knowledge graph from the content you ingest into the platform. We maintain the relationships between your content, and the content sources (i.e. feeds). Optionally, you can enable entity extraction to identify people, places, organizations, products, etc. from your text or transcriptis, and add those to your knowledge graph. You can also enable entity enrichment to connect Wikipedia, Crunchbase or other APIs, and import additional metadata into the knowledge graph.
Do you offer a free trial?
Good question! When you signup to Graphlit, you are on the Free Tier. This gives you full access to all the features of the Graphlit API, with no trial expiration. You are only limited by the amount of content you can ingest into your account. By adding a payment method, you can upgrade to a paid tier, which gives higher quota limits - and you will pay a flat monthly platform fee, and any credit usage.
Can I use this to build a chatbot or copilot?
Definitely! Graphlit provides everything you need to build a RAG application, such as a chatbot or copilot, with our easy-to-use API. Compared to the OpenAI Assistants API, Graphlit handles a wider range of content formats, and offers higher limits on storage capacity. Also, we provide vector-based semantic search as well as RAG conversations. In addition to a small monthly platform fee, you are charged for credit usage based on the volume of content ingested into the platform, LLM tokens used, and other cloud API usage. But you can also bring your own LLM keys, and pay for those tokens yourself.
Can I extract text from PDFs or Word documents?
Yes! You can ingest any type of document, including PDF, Word document, Powerpoint presentation, Markdown, etc. and Graphlit will extract structured text. You can optionally use a content workflow with Azure AI Document Intelligence models, for high-quality OCR text and table extraction. In our testing, we have seen comparable or better results using Azure AI Document Intelligence for OCR, compared to Unstructured.IO, LlamaParse, and other open source PDF extractors.
How is this different than LangChain?
LangChain is a leading frameworks for building LLM applications, but it is not a managed platform. When building RAG applications, you would need to integrate LangChain with LLMs, vector databases, and cloud infrastructure. There are many pieces to assemble to have a production-grade application with LangChain, and you are left to solve these DevOps problems yourself. WIth Graphlit, you can start building your RAG application immediately - no assembly required. We provide a scalable, managed API, where you just have to point us at your unstructured data, and we do the rest. We handle text extraction, text chunking, vector embeddings, as well as RAG conversations (with history). Also, Graphlit handles multi-tenant semantic search out of the box.
How is this different than Unstructured.IO?
Unstructured.IO is focused on the extraction of text from PDFs and other documents, and the partitioning of those documents. They are not a RAG-ready platform, in that they don't directly connect your data to LLMs. They are also not multimodal-ready. They are currently focused on textual formats, not any other media formats. You still need to build an end-to-end unstructured data pipeline around their technology. By using Graphlit, you can start building multimodal RAG applications immediately - no assembly required. We offer best-in-class OCR and text extraction, using Azure AI Document Intelligence. Graphlit also supports a wider range of file formats and metadata than Unstructured.IO, and has built-in audio transcription.
Do I need my own vector database, like Pinecone or Qdrant?
Nope! With Graphlit, you get everything you need to create RAG applications out of the box. We already have integrated best-in-class vector search and vector embeddings into the platform.
Do you support a JSON mode for text extraction?
Yes! Graphlit stores extracted text and tables in our own JSON format, which you can access through our API. You have the option of using Graphlit for ingestion and text extraction only, and then post-processing the JSON format with your own application. But in most cases, you would have Graphlit generate vector embeddings from the extracted text, and handle the RAG conversations for you.
Do you support audio and video transcription?
For sure, we handle a variety of audio and video formats, such as MP4, MP3, WAV and AAC/M4A. Any audio or video file ingested into Graphlit will have audio automatically transcribed using the latest Deepgram speech-to-text model.
Can I build a knowledge graph from my data?
Certainly! Graphlit automatically builds a knowledge graph from the content you ingest into the platform. We maintain the relationships between your content, and the content sources (i.e. feeds). Optionally, you can enable entity extraction to identify people, places, organizations, products, etc. from your text or transcriptis, and add those to your knowledge graph. You can also enable entity enrichment to connect Wikipedia, Crunchbase or other APIs, and import additional metadata into the knowledge graph.
Do you offer a free trial?
Good question! When you signup to Graphlit, you are on the Free Tier. This gives you full access to all the features of the Graphlit API, with no trial expiration. You are only limited by the amount of content you can ingest into your account. By adding a payment method, you can upgrade to a paid tier, which gives higher quota limits - and you will pay a flat monthly platform fee, and any credit usage.