AI / ML / Platform Engineer

Building production-ready AI platforms with a focus on LLM systems, AI agents, and machine learning engineering. London, UK.

About

My journey into AI systems began with a foundation in software development and cloud computing architecture. After completing my MSc in Cloud Computing, I recognized that the next frontier in building scalable, intelligent platforms lay in operationalizing machine learning models and large language systems.

Today, I focus on designing and deploying production AI systems that solve real business problems. My work centers on LLM-based agents, document intelligence platforms, and automated workflows that leverage both pretrained and fine-tuned models. I prioritize system design thinking—ensuring that AI components integrate cleanly with existing infrastructure, handle scale, and maintain reliability.

Beyond implementation, I care deeply about architecture decisions: when to use pretrained models versus fine-tuning, how to structure agent-based systems for maintainability, and how to build AI platforms that are both secure and compliant with regulatory requirements.

Production code I've built remains private under organization accounts, as is standard for proprietary platforms. I share technical insights through architecture documentation, system design case studies, and technical decision records that focus on patterns and principles rather than implementation details.

AI & LLM Systems

Production systems built with LLMs, focusing on architecture decisions, model selection, and operational concerns.

VoxyDocs: Agent-Based Document Intelligence

Problem Statement

Organizations process thousands of documents daily with varying formats and structures. Traditional OCR and template-based extraction fails when document layouts change, creating maintenance overhead and poor accuracy on unstructured content.

Why LLMs

LLMs excel at understanding context, extracting structured information from unstructured text, and handling document variations without rigid templates. Agent-based architectures allow for multi-step reasoning, validation, and error recovery that traditional pipelines cannot provide.

Architecture Overview

Multi-agent system with specialized roles: document parser agent (handles format detection and text extraction), extraction agent (uses LLM to identify key fields), validation agent (cross-checks extracted data against business rules), and enrichment agent (adds metadata and classifications). Agents communicate via a message queue, with a coordinator orchestrating workflow state.

Data Flow

Documents enter via API or batch upload → storage layer (S3/GCS) → parser agent extracts raw text → extraction agent processes text through LLM API with structured output format → validation agent reviews extractions → enrichment agent adds tags/classifications → results stored in database with confidence scores → webhook/callback notifies client system.

Model Choices

Base extraction uses GPT-4-turbo for structured outputs (JSON mode) due to reliability. Fine-tuned Llama 2 13B deployed on GPU instances for high-volume, cost-sensitive workflows. Sentence transformers (all-MiniLM-L6-v2) for semantic search and document similarity. Embeddings stored in vector database (Pinecone) for retrieval-augmented workflows.

Deployment Considerations

API deployed on Kubernetes with auto-scaling based on queue depth. LLM calls rate-limited and cached where possible. Async processing pipeline (Celery/RQ) for batch jobs. Model endpoints abstracted behind a service layer for easy swapping. Monitoring includes latency percentiles, token usage, cost tracking, and accuracy metrics via human-in-the-loop validation samples.

VoxyDocs: Multi-Agent Architecture
Client API/ Batch UploadS3/GCSStorageMessage QueueRedis/RabbitMQCoordinatorOrchestratorParser AgentFormat DetectionText ExtractionExtraction AgentLLM ProcessingStructured OutputValidation AgentBusiness RulesCross-CheckEnrichment AgentMetadataClassificationLLM ServicesGPT-4-turboLlama 2 13BEmbeddingsPineconeVector DBPostgreSQLResults StorageWebhookCallback

Job Application Automation AI Agent

Problem Statement

Applying to multiple positions requires customizing cover letters, extracting relevant experience, and matching qualifications to job descriptions. Manual process is time-intensive and often leads to lower-quality applications due to repetition fatigue.

Why LLMs

LLMs can analyze job descriptions, extract key requirements, match them against candidate profiles, and generate personalized application materials. Agent frameworks enable multi-step workflows: research, drafting, review, and submission with human oversight at critical steps.

Architecture Overview

Single autonomous agent with access to multiple tools: resume parser, job board API client, document generator, and email client. Agent maintains conversation state and can iterate on outputs. Human-in-the-loop checkpoints before final submission. Tool-use pattern (ReAct) for deterministic actions, LLM for reasoning and generation.

Data Flow

User provides resume and job URLs → agent fetches job descriptions → extracts requirements using LLM → parses resume into structured format → agent generates tailored cover letter with LLM → user reviews → agent revises if needed → final documents generated → submission via email/API (if approved). State persisted in database for resume and audit trail.

Model Choices

GPT-4 for reasoning and generation tasks (best quality for personalized content). Claude 3 Opus for complex analysis tasks. OpenAI embeddings (text-embedding-3-small) for semantic matching between resume sections and job requirements. No fine-tuning needed—prompt engineering and RAG techniques provide sufficient personalization.

Deployment Considerations

Agent runs as a long-lived process (or serverless function with state management) with API endpoints for job initiation and status checks. Rate limiting per user to prevent abuse. Output caching for repeated job types. Human approval workflow integrated via webhooks. All LLM interactions logged for compliance and improvement.

Job Application Automation: Agent Architecture
User InputResume + Job URLsPreferencesAutonomous AgentGPT-4 / Claude 3 OpusReAct PatternState ManagementTool UseResume ParserStructured DataExtractionJob Board APIFetch DescriptionsRequirementsDocument GeneratorCover LettersApplicationsEmail ClientSubmissionTrackingEmbeddingstext-embedding-3Semantic MatchingResume ↔ JobsHuman-in-the-LoopReview & ApproveCheckpointState StorageConversation HistoryAudit Trail

Secure Document Verification Workflows

Problem Statement

Financial and legal institutions need to verify document authenticity, extract sensitive information securely, and ensure compliance with regulations (GDPR, KYC). Traditional manual review is slow, expensive, and inconsistent. Automated systems must balance accuracy with privacy and auditability.

Why LLMs

LLMs can extract structured data from identity documents, contracts, and certificates while maintaining high accuracy. Combined with cryptographic verification and zero-knowledge techniques, they enable automated processing while preserving privacy. LLMs can also detect anomalies and flag documents requiring human review.

Architecture Overview

Hybrid system: LLM extracts structured data, cryptographic verification validates document integrity, rule-based engine applies compliance checks. Documents encrypted at rest and in transit. Extraction happens in isolated, audited environment. PII redaction layer before any external API calls. Audit log captures all operations for compliance reporting.

Data Flow

Document uploaded with encryption → stored in secure vault → decryption in isolated processing environment → LLM extraction with PII detection → structured data stored separately from source documents → compliance rules engine evaluates data → results encrypted and returned → audit log updated. Original documents deleted after verification window unless retention required.

Model Choices

GPT-4 with structured outputs for reliable extraction. Local model (Llama 2 fine-tuned) for sensitive PII fields to avoid sending data externally. Named entity recognition models for PII detection. Custom classification model for document type detection. No data sent to external APIs for highest-security workflows; all processing on-premises when required.

Deployment Considerations

Deployed in private cloud or on-premises for sensitive workloads. API endpoints behind authentication and rate limiting. Processing pipeline uses secure enclaves where available. Database encryption at rest. Regular security audits and penetration testing. Compliance reports generated automatically. Fail-closed design: any uncertainty triggers human review rather than auto-approval.

Secure Document Verification: Hybrid Architecture
Document UploadEncryptedTLSSecure VaultEncrypted StorageKMSIsolated ProcessingEnvironmentLLM ExtractionPII DetectionStructured OutputLocal ModelLlama 2 Fine-tunedSensitive PIIGPT-4 APIPII RedactionLayerBefore API CallStructured OutputCompliance RulesEngineGDPR / KYCBusiness RulesCryptographicVerificationDocument IntegrityAuthenticityStructured DataStorageEncrypted DBSeparate from DocsAudit LogAll OperationsCompliance ReportingImmutableEncrypted ResultsReturned to ClientVerification StatusExtracted DataHuman ReviewFlagged CasesFail-ClosedDesign

Full-Stack & Platform Engineering

These capabilities support and enable the AI systems I build. Full-stack engineering, deployment infrastructure, and platform design are foundational to production ML/AI systems.

Web Design & Development

Design decisions focus on accessibility, performance, and user experience. Built responsive interfaces using Next.js, React, and TypeScript with emphasis on semantic HTML, ARIA attributes, and keyboard navigation. Performance optimization through code splitting, image optimization, and efficient state management.

UX decisions prioritize clarity and functionality over decoration. Information architecture supports complex workflows without overwhelming users. Form design, error handling, and feedback loops are engineered for reliability and user trust.

Backend & API Development

Built RESTful and GraphQL APIs using Django, FastAPI, and Node.js. API design emphasizes consistency, versioning, and clear error responses. Authentication and authorization implemented with JWT, OAuth2, and role-based access control.

Async processing pipelines for long-running tasks using Celery, RQ, or serverless functions. Webhook systems for event-driven architectures. Rate limiting, caching strategies, and request validation ensure system reliability and security.

Database Design & Configuration

Designed relational schemas (PostgreSQL) and document stores (MongoDB) based on access patterns and consistency requirements. Optimized queries through indexing strategies, query analysis, and connection pooling.

Vector databases (Pinecone, Weaviate) for semantic search and RAG systems. Redis for caching, session management, and real-time features. Database migrations, backup strategies, and replication configured for production reliability.

Deployment & Infrastructure

Deployed applications on Cloudflare Pages, Vercel, AWS, and Kubernetes. CI/CD pipelines configured with GitHub Actions for automated testing, building, and deployment. Infrastructure as code using Docker, Kubernetes manifests, and Terraform.

Monitoring and observability with logging, metrics, and distributed tracing. Error tracking, performance monitoring, and alerting configured for production systems. Security hardening through environment variables, secrets management, and network policies.

Technical SEO & Analytics

Implemented technical SEO for production platforms: structured data (JSON-LD), sitemap generation, robots.txt configuration, and meta tag optimization. Performance optimization for Core Web Vitals and search engine indexing.

Analytics integration for user behavior tracking, conversion funnels, and performance metrics. A/B testing frameworks for data-driven iteration. Search console monitoring and indexing status tracking.

Audience & Digital Growth Systems

Designed and executed content distribution systems reaching 90K+ users across platforms through data-driven iteration and platform-specific optimization. This work demonstrates systems thinking applied to audience growth, content automation, and analytics-driven product development.

Content System Architecture

Built content production and distribution workflows that maintain consistency across YouTube and Instagram. Automated scheduling, cross-platform publishing, and analytics aggregation enable data-driven iteration without manual overhead.

Content templates and production pipelines reduce time-to-publish while maintaining quality standards. Version control for content assets, metadata management, and performance tracking integrated into the workflow.

Platform Optimization

YouTube: Optimized for algorithm performance through metadata engineering, thumbnail A/B testing, and engagement pattern analysis. Analyzed watch time, retention curves, and click-through rates to iterate on content structure and presentation.

Instagram: Algorithm-aware posting schedules, hashtag strategy based on reach analytics, and content format optimization (Reels, Posts, Stories) driven by engagement metrics. Cross-posting workflows and analytics aggregation for unified performance tracking.

Analytics & Iteration

Built dashboards aggregating metrics across platforms: subscriber growth, engagement rates, content performance, and audience demographics. Identified patterns in high-performing content and applied learnings to production workflows.

A/B testing frameworks for thumbnails, titles, and content formats. Cohort analysis for audience retention and growth attribution. Data-driven decisions on content topics, posting frequency, and platform-specific optimizations.

Scale & Automation

Scaled content production from zero to 90K+ combined audience through systematic iteration. Automation tools for content scheduling, cross-platform publishing, and analytics collection reduce operational overhead while maintaining quality.

Content primarily in Tamil for regional audience, demonstrating ability to build systems that serve specific market segments. Growth achieved through consistent execution, platform optimization, and data-driven iteration rather than viral moments.

Note: This work is framed as systems engineering and growth platform development, not influencer marketing. The focus is on building scalable content distribution systems, analytics infrastructure, and data-driven iteration processes.

Case Studies

Production systems built and deployed, with emphasis on architecture, system design, and operational considerations. Detailed case studies are available in the architecture documentation repository.

VoxyDocs

AI document intelligence platform for automated data extraction and processing

Problem

Organizations struggle with extracting structured data from diverse document formats. Manual data entry is slow and error-prone, while template-based automation breaks when document layouts change. This creates bottlenecks in processes like invoice processing, contract analysis, and form handling.

Constraints

Must handle document format variations without manual template updates. System must scale horizontally for batch processing. Cost per extraction must remain reasonable at scale. Extraction accuracy must meet business requirements with confidence scoring. Integration must work with existing enterprise systems via APIs.

Architecture

Multi-agent system with specialized roles: document parser agent (handles format detection and text extraction), extraction agent (uses LLM to identify key fields), validation agent (cross-checks extracted data against business rules), and enrichment agent (adds metadata and classifications). Agents communicate via message queue (Redis/Celery), with a coordinator orchestrating workflow state. Microservices architecture allows independent scaling of components. Async processing pipelines handle batch jobs while maintaining low latency for real-time requests.

Stack

PythonFastAPIOpenAI GPT-4LangChainPostgreSQLRedisCeleryDockerKubernetesAWS S3Pinecone

Deployment

API deployed on Kubernetes with auto-scaling based on queue depth. LLM calls rate-limited and cached where possible. Async processing pipeline (Celery) for batch jobs. Model endpoints abstracted behind a service layer for easy swapping. Monitoring includes latency percentiles, token usage, cost tracking, and accuracy metrics via human-in-the-loop validation samples. CI/CD pipeline with automated testing and blue-green deployments.

Learnings

Agent-based architectures require careful state management and error handling. Learned to balance between autonomous agents and deterministic validation steps. Cost optimization became critical—implemented caching, batch processing, and model routing (using cheaper models for simple extractions, expensive ones only when needed). Found that structured outputs (JSON mode) dramatically improved reliability over free-form extraction. Human-in-the-loop workflows are essential for high-stakes extractions.

VoxyDocs: Multi-Agent Architecture
Client API/ Batch UploadS3/GCSStorageMessage QueueRedis/RabbitMQCoordinatorOrchestratorParser AgentFormat DetectionText ExtractionExtraction AgentLLM ProcessingStructured OutputValidation AgentBusiness RulesCross-CheckEnrichment AgentMetadataClassificationLLM ServicesGPT-4-turboLlama 2 13BEmbeddingsPineconeVector DBPostgreSQLResults StorageWebhookCallback

Verilett

Secure rental and payments platform with automation workflows

Visit →

Problem

Rental property management involves complex workflows: tenant verification, contract generation, payment processing, and compliance tracking. Manual processes lead to delays, errors, and security risks. Property managers need automation while maintaining trust and legal compliance.

Constraints

Must handle sensitive financial and personal data with strict security requirements. System must comply with GDPR and financial regulations. Document verification must be accurate and auditable. Payment processing must integrate with Stripe while maintaining reconciliation workflows. User experience must balance automation with transparency and control.

Architecture

Full-stack platform with Next.js frontend and Node.js backend. Document verification system extracts and validates tenant information from IDs and financial documents using LLM-based extraction with validation rules. Automated contract generation personalizes standard templates based on property and tenant data using LLM APIs. Payment processing integrated via Stripe API with webhook handlers for event-driven workflows. Database schema designed for audit trails and compliance reporting. Email automation via AWS SES for notifications and reminders.

Stack

Next.jsTypeScriptNode.jsPostgreSQLStripe APIOpenAI APIAWS SESDockerVercelPrisma

Deployment

Frontend deployed on Vercel with edge functions for API routes. Database hosted on managed PostgreSQL service with automated backups. Docker containers for local development and testing. CI/CD pipeline with automated testing and staging environments. Environment variables and secrets managed through Vercel and AWS parameter store. Monitoring and error tracking configured for production.

SEO / Growth

Implemented technical SEO: structured data for property listings, sitemap generation, and meta tag optimization. Performance optimization for Core Web Vitals. Analytics integration for user behavior tracking and conversion funnels. Search console monitoring for indexing status.

Learnings

Security and compliance are non-negotiable when handling financial and personal data. Learned to implement proper encryption, access controls, and audit logging. Document verification accuracy required careful prompt engineering and validation rules. Payment automation needed robust error handling and reconciliation workflows. User experience matters even for B2B tools—automation should feel helpful, not opaque. Audit trails are essential for compliance and debugging.

Verilett: Rental Platform Architecture
Next.js FrontendTypeScriptReactVercelNode.js APIREST APIAuthenticationAuthorizationDocument VerificationID ExtractionLLM ProcessingValidationContract GenerationOpenAI APITemplate PersonalizationLLM-basedPostgreSQLPrisma ORMData StorageEncryptedStripe APIPayment ProcessingSchedulingRemindersAWS SESEmail NotificationsRemindersAlertsWorkflow EngineAutomationState ManagementOrchestrationAudit LogComplianceTracking

RoomTo.Live

Real-world product platform connecting property seekers with verified listings

Visit →

Problem

Property search platforms suffer from low-quality listings, fake listings, and poor matching between seekers and properties. Users waste time on irrelevant results or unreliable landlords. The platform needs to ensure listing quality and provide intelligent matching without being intrusive.

Constraints

Must balance listing verification rigor with user friction—too much verification reduces listings, too little reduces trust. Matching algorithm must handle high concurrency during peak search times. System must scale horizontally as user base grows. Real-time updates (new listings, availability changes) must propagate quickly. Cost per search must remain reasonable despite using embeddings and LLM APIs.

Architecture

Marketplace platform with Next.js frontend and Node.js backend. AI-powered listing verification uses LLMs to analyze descriptions and flag inconsistencies. Recommendation engine combines traditional filters with semantic search (OpenAI embeddings) stored in vector database. Elasticsearch for full-text search and filtering. Event-driven architecture handles real-time updates via message queues. WebSocket connections for live notifications. Redis for caching search results and session management. Image storage on S3 with CDN delivery.

Stack

ReactNext.jsNode.jsPostgreSQLElasticsearchOpenAI EmbeddingsRedisWebSocketsAWS LambdaS3

Deployment

Frontend deployed on Cloudflare Pages with edge caching. Backend API on AWS Lambda with API Gateway. Database on managed PostgreSQL with read replicas. Elasticsearch cluster for search functionality. Redis cluster for caching. CI/CD pipeline with automated testing. Monitoring and alerting configured for all services. Image optimization and CDN delivery for performance.

SEO / Growth

Technical SEO implemented: structured data for property listings (Schema.org), sitemap generation, and meta tag optimization. Performance optimization for Core Web Vitals. Analytics integration for search behavior and conversion tracking. Search console monitoring for indexing and performance metrics.

Learnings

Recommendation systems benefit from hybrid approaches—combining collaborative filtering, content-based filtering, and LLM-powered semantic understanding. Learned that user feedback loops are essential for improving recommendations. Handling real-time updates (new listings, availability changes) required careful event sourcing patterns. Marketplace platforms need to balance trust (verification) with friction (too much verification reduces listings). Caching strategies are critical for cost optimization with embedding APIs.

RoomTo.Live: Marketplace Architecture
React FrontendNext.jsReal-time UIWebSocketsAPI GatewayNode.jsREST APIWebSocket ServerListing VerificationLLM AnalysisQuality ChecksInconsistency DetectionRecommendationEngineHybrid ApproachSemantic SearchEmbeddingsCollaborative FilterEmbeddingsOpenAItext-embedding-3Semantic MatchingPostgreSQLUser DataListingsMetadataElasticsearchFull-text SearchFilteringIndexingRedis CacheQuery ResultsSession DataEvent StreamReal-time UpdatesNew ListingsAvailabilityAWS LambdaAsync ProcessingBackground JobsS3 StorageImagesDocuments

Architecture Documentation

Production systems and proprietary platforms I've built are private under organization accounts. I maintain a public repository of architecture notes, system design documentation, and technical decision records.

System Architecture Notes

A collection of architecture documentation from production AI/LLM systems, including design patterns, decision frameworks, and system design case studies.

Architecture PatternsSystem DesignDecision RecordsCase Studies

Contents Include

  • • Multi-agent architecture patterns
  • • LLM system scaling strategies
  • • Model selection frameworks
  • • Data pipeline architectures
  • • Infrastructure deployment patterns
  • • Security & compliance approaches

Format

  • • Architecture diagrams and schemas
  • • Technical decision records (ADRs)
  • • System design case studies
  • • Trade-off analysis and constraints
  • • Production considerations
  • • Failure cases and lessons learned

Note: All documentation is sanitized and contains no proprietary code, customer data, or sensitive implementation details. Focus is on architectural patterns, decision-making frameworks, and system design principles applicable across production ML/AI systems.

Contact

Interested in discussing AI systems, LLM engineering, or potential opportunities? Get in touch.

Or reach out directly at: pradeepannagarasu@gmail.com