About Skills Experience Projects Contact
Available for opportunities

Utkarsh
Adhran

Software Developer

6+ years building data engineering pipelines, automation systems, and backend solutions. Specialized in large-scale financial market data — processed 1.5TB+ of options & futures data using DuckDB, Parquet, and Python.

6+
Years Experience
1.5TB
Data Processed
4+
Companies
10+
Notable Projects

About Me

I'm a Software Developer based in Noida, India with 6+ years of experience across data engineering, automation, and backend systems. I started my career building RPA bots and automation scripts, and have since moved into large-scale financial data pipelines.

Currently at Open Futures and Commodities, I work with NSE and TAIFEX market data — converting raw datasets into optimized Parquet files using DuckDB, building quantitative trading strategies, and setting up real-time visualization pipelines with QuestDB and Grafana.

I'm passionate about turning raw, messy data into fast, queryable systems and solving bottlenecks through practical engineering. I hold an AWS Certified Cloud Practitioner certification and have hands-on experience with S3 integration and cloud migrations.

📊

Data Engineering

1.5TB+ raw market data pipelines using DuckDB, Parquet & Zstd compression

⚙️

Automation & Scripting

End-to-end RPA bots, PDF parsing, OCR, Selenium automation for enterprise clients

📈

Live Visualization

Real-time market data dashboards via QuestDB + Grafana with rsync automation

☁️

Cloud & Backend

AWS S3 integration, GCP → AWS migration, Django CRM, REST API development

Skills & Tools

Languages
Python Shell Scripting SQL
Data & Databases
Pandas DuckDB MySQL QuestDB PyArrow Dask Parquet Zstd
Web & Visualization
Django Streamlit Grafana React Native
Cloud & DevOps
AWS Linux Git rsync
Automation
Selenium Joblib PyPDF2 OCR / Tesseract smtplib
Certifications
AWS Cloud Practitioner

Experience

💹
Software Developer
Jul 2025 — Present
Open Futures and Commodities Private Limited · Noida
  • Processed NSE market data with Zstd compression & Parquet format via DuckDB — major gains in query performance
  • Converted 1.5TB of raw TAIFEX Options & Futures .dat data into structured Parquet files for quantitative analysis
  • Built Streamlit apps for interactive strategy analysis and testing on large-scale options data
  • Set up live QuestDB + Grafana data visualization pipeline with rsync-based daily scheduling
  • Implemented Joblib caching to reduce repeated computation time across data pipelines
Python DuckDB QuestDB Grafana Streamlit Parquet Zstd Joblib Linux
🔧
Consultant
Oct 2021 — Apr 2025
Genpact · Noida
  • Debugged and tested applications to enhance system performance across various environments
  • Built ML-based data completion POC (Credit Suisse) — rule-based logic for auto-filling incomplete datasets
  • Contributed to RHEL 5 → 8 migration for PIMCO — utility scripts, dependency resolution, application debugging
  • Created automation scripts reducing manual overhead across repeated operational tasks
Python Linux (RHEL) ML / scikit-learn Bash SQL
🌱
Software Developer
Jan 2021 — Oct 2021
FreshoKartz Agri Products Pvt. Ltd. · Jaipur
  • Maintained and enhanced Django/MySQL CRM platform covering sales, inventory & kiosk operations
  • Built React Native mobile apps — employee attendance tracking + farmer ordering platform
  • Contributed to GCP → AWS database migration improving system scalability
Django MySQL Python React Native AWS GCP
🤖
Python Developer
Dec 2019 — Oct 2020
SequelString Solutions And Consultancy Pvt Ltd · Noida
  • Built automation bots for Hero MotoCorp (invoice parsing), Indus Towers (bill extraction), KPMG (ITSM ticketing)
  • Created RPA bot for Aknamed to auto-generate sales orders via Selenium-driven browser automation
  • Automated IndiaMART tender scraping and PDF metadata extraction
Python Selenium Pandas MySQL PyPDF2 OCR

Notable Projects

📦
NSE Market Data Pipeline
Large-scale NSE market data handling — CSV to Parquet conversion via DuckDB with custom filters, Zstd compression, and Joblib caching for repeated computations.
Storage optimization · Query performance
Python DuckDB Parquet Zstd Joblib
🗃️
TAIFEX Options & Futures Pipeline
End-to-end pipeline for 1.5TB of raw TAIFEX .dat data — conversion to structured Parquet, quantitative strategy development, and Streamlit apps for interactive analysis.
1.5TB raw data · Quant strategies
Python DuckDB Streamlit Parquet
📡
Live Market Data Visualization
Real-time data pipeline integrating QuestDB and Grafana for live market visualization. Automated rsync-based scheduling for daily raw market data across different expiries.
Real-time · Automated pipeline
QuestDB Grafana rsync Python
🧠
ML-Based Data Completion (Credit Suisse)
POC evaluating ML algorithms for auto-filling incomplete datasets. Integrated rule-based logic to test multiple scenarios, assessing feasibility for reducing manual processing.
Proof of Concept · Genpact
Python Pandas scikit-learn SQL
🛒
FreshoKartz CRM System
Full-stack CRM covering sales, inventory, vendor purchases, and kiosk operations. Led maintenance, feature development, and deployment cycles on Django + MySQL stack.
Production CRM · End-to-end
Django MySQL Python
🤖
ITSM Workflow Automation (KPMG)
Automated ITSM ticket lifecycle — creation, assignment, and closure using Python and Selenium, with SQL-based logging and Pandas reporting.
Enterprise automation · KPMG
Python Selenium Pandas SQL

Certification

☁️

AWS Certified Cloud Practitioner

Foundational level — Amazon Web Services · Issued Aug 2023

Valid until Aug 2026 Q8WJVL7CWJ41Q5G8

Get In Touch

Let's work together

Open to full-time roles, freelance projects, and interesting data engineering challenges. Based in Noida — open to remote.