S3 Document Pipeline
Auto-process documents uploaded to S3 with text extraction, classification, and routing. Handles PDFs, images, and text files with retries.
connic init my-project --templates=s3-document-pipelineOverview
Triggered by file uploads to an S3 bucket. The intake agent extracts text, metadata, and entities from PDFs, images, and text files. Middleware rejects unsupported file types with StopProcessing before the LLM runs. The classifier assigns categories, confidence scores, and routing destinations. Retries on transient failures.
Use cases
Document management
Automatically classify and route documents to the right team as they land in your S3 inbox.
Contract processing
Extract key entities (parties, dates, amounts) from uploaded contracts and flag for legal review.
Multi-format intake
Process PDFs, images, Word documents, and text files through a single unified pipeline.
Architecture
Scaffolded project structure
Running connic init my-project --templates=s3-document-pipeline creates this file tree.
s3-document-pipeline/
agents/
document-intake.yaml
document-classifier.yaml
document-pipeline.yaml
middleware/
document-intake.py
schemas/
intake-result.json
classification-result.json
requirements.txt
README.mdGet started
Install the template, create a Connic project, and deploy. Choose Git (automatic on push) or CLI (works with any provider).
Prerequisites
- Python 3.10+
- A Connic account (create a project first)
- API key for your LLM provider (e.g. Gemini, OpenAI) to add in project variables
Install and scaffold
Install the SDK and create a project from this template.
pip install connic-composer-sdkconnic init my-project --templates=s3-document-pipelineThen cd my-project
Deploy
Pick your deployment method. Git auto-deploys on push; CLI works with GitLab, Bitbucket, or no Git.
Git integration
- In Connic: Project Settings → Git Repository, connect your GitHub repo
- Settings → Environments: map branch (e.g.
main) to Production - Push your scaffolded project to that repo
git add .
git commit -m "Add S3 Document Pipeline template"
git push origin mainCLI deploy
- In Connic: Project Settings → CLI, create an API key and copy project ID
- Run
connic loginin your project folder connic testto try with hot-reload, orconnic deployfor production
connic login
connic test # Ephemeral dev env with hot-reload
connic deploy # Deploy to productionConnect and configure
Add an S3 inbound connector with your bucket name and AWS credentials. Configure S3 Event Notifications to point to the connector's webhook URL. Add your LLM provider API key in Project Settings → LLM Provider API Keys.
Template source
Browse the full template, contribute improvements, or fork for your own use.
connic-org/connic-awesome-agents/tree/main/s3-document-pipelineExplore other templates
View allTelegram Personal Assistant
Multimodal Telegram assistant that processes text, images, voice, video, and documents. Searches the web, saves notes and knowledge, and schedules follow-up reminders with persistent sessions.
Lead Enricher
Enrich and qualify new signups by researching their company and scoring fit against your ideal customer profile. Saves qualified leads.
Compliance Auditor
Scheduled compliance scanning combining web search for regulatory updates, MCP docs, and knowledge base history. Produces scored summaries.
Email Helpdesk
End-to-end email support with IMAP polling, auto-reply filtering, and intent classification. Drafts contextual replies from your knowledge base.