Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings
View Saicharansid's full-sized avatar
🎯
Focusing
🎯
Focusing

Block or report Saicharansid

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Saicharansid/README.md

Hi, I’m Sai Charan — Data Engineer ⚙️ | FinTech ⚡ | AI/ML Curious 🤖

I build reliable, cloud-native data pipelines that turn messy, high-volume data into trustworthy, analytics-ready assets. My sweet spot is the FinTech edge where fraud, risk, and growth analytics meet streaming & batch pipelines, Delta/Medallion architectures, and governed data products.

  • 🧩 Core: Python, SQL, PySpark, Airflow, dbt, Kafka, Terraform
  • ☁️ Cloud: AWS (Glue, S3, Lambda, Redshift), Azure (ADF, Synapse), GCP (BigQuery)
  • 🗄️ Warehousing/Lakehouse: Snowflake, Databricks, Delta Lake, Redshift
  • 🔐 Data Ops & Governance: GitHub Actions, CI/CD, lineage/catalog, quality gates
  • 🧠 AI/ML-adjacent: feature pipelines, model serving hooks, vector ETL, RAG plumbing

What I’m focused on right now

  • FinTech pipelines: fraud/risk event ingestion, CDC from OLTP, real-time features for models
  • Automation-first data ops: infra-as-code, CI/CD for ELT, unit/data tests, SLAs/SLOs
  • AI/ML integration: feature stores, batch/stream feature joins, embedding & retrieval pipelines

Selected Projects (recruiter highlights)

  • Regulatory Compliance Lakehouse (AWS + Delta + Redshift Spectrum)
    Consolidated trade/audit/regulatory data to a governed lakehouse with versioned, queryable history and sub-minute analytics. Result: 2× faster compliance reporting and end-to-end traceability.

  • Omnichannel Retail Analytics (Glue + Kafka + Athena)
    Real-time ingestion + curated marts powering ROAS and attribution. Result: +35% attribution accuracy and −70% manual reconciliation.

  • Reddit → AWS End-to-End Pipeline (Airflow + S3 + Lambda + Quicksight)
    Automated ingestion, bronze/silver/gold transforms, and scheduled dashboards. Designed for easy extension into feature engineering and LLM/RAG demos.

Want the technical deep-dive? See my portfolio: https://sid-data-portfolio.lovable.app/


How I work

  • Design for observability: metrics, data quality tests (schema/constraints/freshness), lineage
  • Treat data as a product: clear contracts, SLAs, versioning, documentation
  • Automation by default: reproducible environments, templated pipelines, IaC

Tech I enjoy

Python · SQL · PySpark · Airflow · dbt · Kafka · Databricks · Delta Lake · Snowflake · Redshift
AWS Glue · S3 · Lambda · Azure Data Factory · Synapse · BigQuery
Terraform · GitHub Actions · Power BI · Tableau


Let’s collaborate

Open to FinTech data engineering, feature/ML data pipelines, governed lakehouse builds, and AI-powered analytics integrations.

If you’re a recruiter: happy to share code walk-throughs, architecture diagrams, and before/after performance metrics.

Pinned Loading

  1. saicharansid.github.io saicharansid.github.io Public

    Here's my resume for reference

    HTML

Morty Proxy This is a proxified and sanitized view of the page, visit original site.