























01 // VECTOR INGESTION
Tech: Pinecone, OpenAI, Vectors
02 // PROTOCOL-BASED AGENTS
Tech: LangChain, Tools, JSON
03 // DYNAMIC LOGGING
Tech: Google APIs, Webhooks, HTTP
The Challenge: AI models are useless if their data is old. The client needed a way to "teach" the AI new policies instantly by just dropping a PDF into a folder.
The Architecture: A webhook-triggered flow that lists files from a Google Drive Inbox. Custom JS nodes convert Docs to text while handling PDFs as binaries. Before upserting, we delete old vectors to prevent "Ghost Data".
"I don't just copy files. I manage Vector Lifecycles."
if (!item.binary && item.json.data) { // Handle Google Doc Export const textContent = item.json.data; item.binary = {; data: Buffer.from(textContent).toString('base64'), mimeType: 'text/plain', fileName: original.name + '.txt' }; } else { // Handle Native Binary (PDF/IMG) item.binary.data.fileName = original.name; } return $input.all();
The Challenge: Standard chatbots hallucinate. The client needed a bot that acted like a trained Sales Rep.
The Architecture: A LangChain Agent with 4 distinct Protocols (Consult, Product Search, Info, Closing). The agent autonomously decides when to query the Inventory DB or capture lead data into Sheets.
The Challenge: The best freelance jobs are gone in minutes. Manually refreshing feeds is a losing game. I needed a system to beat standard notifications and draft 6-figure proposals instantly.
The Architecture: A real-time monitor that bypasses RSS for direct GraphQL access. It filters for high-ticket clients (Payment Verified, Spend > $500), then triggers a "Chief Engineer" agent to estimate the work and a "Sales Agent" to write the pitch.
The Challenge: A client sat on 10,000+ scanned legacy contracts. They were just images—"Dark Data" invisible to search and unreadable by LLMs. Standard tools crashed on the corrupted file headers.
The Architecture: A brutalist Python pipeline. I use PyMuPDF for bit-level surgery to fix broken XREF tables, then pipe the raw image streams into Tesseract to reconstruct a hidden text layer.
**Status:** Active
**Summary:** Binding agreement between...
The Challenge: A legal firm had 50,000+ legacy MS Word documents from the 2000s. They were riddled with XML corruption that caused standard Python parsers to panic. The mission? Convert them all to clean Markdown for a RAG knowledge base.
The Architecture: A fault-tolerant conversion pipeline. I bypassed the standard python-docx library and went straight for the XML guts using Pandoc with custom Lua filters to perform "tree surgery" on broken tags.
The Challenge: Clients needed to query internal PDFs quickly, but standard AI tools hallucinated facts and couldn't access private files. They needed a "Corporate Brain" that only spoke the truth.
The Architecture: A precision RAG pipeline. I orchestrated LangChain to chunk thousands of documents and index them in ChromaDB. When a user asks a question, the system retrieves the exact source paragraphs and forces GPT-4o to answer only using that evidence.
Production workflows and interfaces running in the wild.

