Rabindra Kharel
π¨βπ» Staff Engineer β Data, Analytics & AI Platforms
$ whoami
Staff Engineer β Backend/API Platforms β’ Data Engineering β’ AI/ML Infrastructure
Experience: 10+ years architecting fault-tolerant, production-scale data systems
π Professional Summary
Highly experienced Staff Engineer specializing in Backend/API Platforms, Data Engineering, and AI/ML Infrastructure. Adept at designing fault-tolerant, cost-aware architecture that transforms complex data, analytics, and ML requirements into scalable, observable, production systems. Proven track record translating complex requirements into robust data engineering architecture, data models, and analytic solutions. Experienced in leading cross-functional teams to deliver end-to-end Analytics and AI solutions that unlock insights and drive business value.
πΌ Professional Experience
- Fault-Tolerant Ingestion: Built exactly-once ingestion pipeline streaming PB-scale data from 50+ sources into S3/Iceberg lakehouse (AWS Lake Formation)
- Data Lake Architecture: Architected both batch and streaming pipelines using Kafka, Flink, SQS, Lambda, and S3
- Data Mesh Implementation: Led end-to-end architecture and provisioning for data mesh and data-product solutions for Alight Commercial Data Lake
- Security & Governance: Integrated AWS Identity Center; authored tag-driven policies enforcing RBAC, RLS, and dynamic PII masking
- Performance Optimization: Tuned Redshift WLM, Short Query Acceleration, and Query Monitoring Rulesβdropped BI latency to <3s while trimming monthly cost 30%
- CI/CD Pipeline: Engineered GitHub-based CI/CD pipelines for database objects and ETL code deployment
- QuickSight Integration: Wired curated layers to QuickSight and downstream systems driving actionable insights
- Lifecycle Management: Developed lifecycle/compaction jobs and hot/warm/cold tiering; piped lineage & metrics to CloudWatch for single-pane SLO dashboards
- Advanced Analytics: Data Vault 2.0, clustering keys for collocated joins, evaluating pruning stats & disk spills, Dynamic Tables/MV for precomputed reports
- Real-time & Batch Pipelines: Engineered pipelines ingesting >3TB/day from eCommerce, clickstream, POS, marketing channels, and social media
- ML Pipeline Development: Built ML pipelines including data preprocessing, feature engineering, data quality checks, batch training, and ModelOps for batch/real-time training
- Identity Resolution: Built customer identity-resolution, context resolution, de-duplication algorithms, and "anonymous-to-known" modules, boosting match rate
- Data Modeling & Optimization: Designed conceptual, logical, and physical data models in Redshiftβoptimizing denormalization, dist/sort keys, compression encodings, automated analyze/vacuum, improving average query latency 40%
- Query Optimization: Assessed business reports, query patterns, and ELT loads to identify optimal Distribution Key, Sort Key, and encoding for each table and column
- Marketing Activation: Integrated analytics-ready data into activation systems like Google AdWords, Mailchimp, and Facebook Marketing for closed-loop marketing
- Cost Optimization: Drove workload right-sizing, reducing EC2 spend 28%
- ML CI/CD Pipelines: Implemented ML CI/CD pipelines covering automated training, evaluation, drift monitoring, canary releases, and auto-promotion
- AI/ML Workflow Automation: Designed and implemented CI/CD pipelines for AI/ML workflows including model training, prediction, and deployment automation using GitHub
- Data Pipelines: Engineered batch data training pipelines and real-time streaming prediction systems using Python
- Model Serving: Served latency-critical XGBoost & deep-learning models on GPU and EKS with gRPC/REST; C++/Rust inference
- Feature Engineering: Developed and deployed feature pipelines for preprocessing and serving data for machine learning models
- Kubernetes Deployment: Automated model deployment processes leveraging Kubernetes and Amazon EKS for scalable, containerized applications
- Artifact Management: Managed artifacts & lineage in S3/MLflow; added SMOTE and focal-loss strategies plus concept-drift alerts
- Containerization: Deployed containerized ML workflows using Docker and Kubernetes for high availability and scalability
- Data Warehouse Design: Designed Data Warehouse & BI Analytics stack, ingesting omni-channel commerce data from disaggregated sources (eCommerce, Brick & Mortar sales, Oracle Pricing Systems, Oracle Customer CX)
- ETL Solutions: Designed data warehousing solutions leveraging ETL workflows for eCommerce and retail data
- Team Leadership: Led 8 onshore/offshore engineers, writing specs, test plans, and deployment playbooks
- Unified Analytics: Combined POS, RMS, RPM, and RESA feeds into Unified Retail Analytics and Customer analytics
- Requirements Analysis: Evaluated business requirements, devised adaptable data engineering frameworks, directed development teams, and ensured prompt completion of technical assignments
- Data Modeling: Worked as offshore BI developer to build data models (Star Schema and Dimensional Modeling), reports and dashboards
- ODI ETL Development: Developed ODI ETL jobs integrating OLTP sources (POS, RMS, RPM, RESA) into Oracle Retail Insight
- Business Collaboration: Collaborated with business leads to understand data and business requirements to build impactful and insightful analytic dashboards
- Automation: Automated nightly loads & validation scriptsβcut manual QA effort 50%, raised SLA adherence to 99.7%
- Integration Development: Developed integration code for OLTP systems into data warehouse
- Community Engagement: Organized hackathons and hosted community events to promote W3C web standards
- Developer Advocacy: Engaged with developer community to advocate for web accessibility and open web principles
- Workshop Facilitation: Facilitated workshops to encourage adoption of modern web technologies and standards
- Recognition: Excellence Award recipient at Opera Software
π οΈ Technical Stack
πΎ Storage & Lakehouse
S3 β’ Iceberg Tables β’ Snowflake β’ Redshift β’ MongoDB β’ DynamoDB
π Streaming & Processing
Kafka β’ Flink β’ AWS Lambda β’ Glue β’ EventBridge β’ SQS FIFO β’ Kinesis
βοΈ Processing & ETL
EMR (Spark) β’ Flink β’ SageMaker Processing β’ Glue ETL
π Orchestration
Airflow β’ DBT β’ Talend β’ IICS
π DevOps & Infrastructure
Terraform β’ GitHub Workflows β’ Docker β’ Kubernetes β’ EKS β’ CloudFormation
π» Programming
Python β’ Java β’ SQL β’ Unix Scripting β’ Advanced SQL
π BI & Analytics
Looker β’ QuickSight β’ Athena
π€ AI/ML Platforms
AWS SageMaker β’ MLflow β’ Model Registry β’ XGBoost
π Data Governance
AWS Lake Formation β’ IAM β’ Identity Center β’ Data Masking β’ RBAC/RLS/CLS
π― Core Expertise
βββ Data Platform & API Engineering
β βββ High-throughput ingestion pipelines
β βββ Massively parallel/distributed compute
β βββ Backend API stacks (serverless + microservices)
β
βββ Data Modeling & Architecture
β βββ Data Vault 2.0 & Dimensional Modeling
β βββ SCD behaviors in columnar/distributed databases
β βββ Data Lake design (S3/Iceberg Lakehouse)
β βββ Clustering, collocated joins, query performance tuning
β βββ Medallion architecture patterns
β
βββ Streaming & Batch Ingestion
β βββ Lambda/Kappa architectures into AWS data lakes
β βββ Real-time and batch processing pipelines
β βββ Distributed architectures (Kafka, Kinesis, Flink)
β
βββ Performance & Cost Optimization
β βββ Query-plan forensics & compute right-sizing
β βββ Hot/warm/cold tiering & compaction
β βββ Dist/sort keys & collocated joins
β βββ Scaling policies & resource monitors
β βββ Result caching & multi-cluster setup
β
βββ Data Governance & Security
β βββ Policy as code for governance
β βββ Dynamic data masking & PII controls
β βββ Tag-based subscriptions & RBAC/RLS/CLS
β βββ Schema evolution & registry management
β βββ Data contracts & single sign-on
β βββ Data subscription frameworks
β
βββ DevOps / BIOps / MLOps
βββ Terraform-driven infrastructure
βββ GitHub Actions CI/CD
βββ Event-driven workflows (Kafka, Flink, Lambda)
βββ Reproducible ML pipelines & model deployment
π Education
| Degree |
Institution |
Year |
| Bachelor of Engineering - Computer Science & Engineering (Honours) |
Tribhuvan University β Institute of Engineering, Pulchowk Campus, Nepal |
2008 - 2012 |
π Certifications
| Certification |
Issuer |
Validity |
Credential ID |
| AWS Certified Data Analytics β Specialty |
Amazon Web Services |
Feb 2023 - Feb 2026 |
NNV34CPC6J1Q1C39 |
| AWS Certified Cloud Practitioner |
Amazon Web Services |
Jan 2023 - Jan 2026 |
NNV34CPC6J1QC39 |
π Honors & Awards
- Winner β NASA Space Apps Hackathon (Software Category) - National-level winner for Spatio-temporal Data Anomaly Detection (2013)
- Winner β World Bank Open Data Hackathon - Created map visualization of open aid flow data (2013)
- Excellence Award β Opera Software - Recognition for outstanding contribution to web standards evangelism