💻 Data Engineer | Snowflake | ELT / ETL | AI-Driven Data Solutions
Snowflake Certified - Data Engineering Professional Certificate
📍 India | 🌎 Open to roles across India (OnSite & Remote) and abroad (Remote) | ⚡ Immediately Available
I design and build data pipelines, cloud warehouses, and AI-driven analytics solutions that help businesses move from raw data to reliable decisions.
Over 3 years, I've worked across two ends of the data stack - engineering pipelines in Snowflake for a real client project (Mercedes-Benz: USA & Canada) at Infosys, and owning the full BI function at Troy Consultancy where I built dashboards used for daily decision-making.
My focus areas:
- ELT pipeline development in Snowflake - from raw ingestion through Bronze → Silver → Gold layers
- Automation using Snowpipe, Streams, and Tasks for CDC and scheduled workflows
- AI-driven data engineering using Snowflake Cortex (Code & Analyst) for assisted SQL development and natural language querying
- Semantic layer development to enable business-friendly data access
- BI & reporting with Power BI, connecting to SQL Server, Excel, and web sources
- Query & warehouse optimization - focused on performance and cost efficiency
Built an end-to-end Snowflake data platform integrating AI-assisted development and semantic analytics. Implemented Medallion Architecture (Bronze → Silver → Gold) to process 13+ source files into an analytical Star Schema with SCD Type 2. Leveraged Snowflake Cortex Code for AI-assisted SQL generation, query optimization, and pipeline development, reducing manual effort. Built a semantic model using Cortex Analyst to enable natural language-to-SQL querying on 86K+ sales records, allowing business users to interact with data without writing SQL.
Full enterprise-style data warehouse built using Microsoft SQL Server, implementing a Bronze → Silver → Gold layered architecture with CRM and ERP source integration, stored procedure-based ETL, star schema modeling, and a Sales Data Mart with dim_customers, dim_products, and fact_sales.
Production-style Snowflake pipeline modeled on a food delivery platform, covering initial & delta loads, CDC using Streams, SCD Type 2 dimensions, a star schema fact table at order-item granularity, data governance with Tags & Masking Policies, and full automation via Stored Procedures and Tasks.
Enterprise-scale retail analytics solution for a 5M+ customer ecommerce company spanning 15 countries. Built on Snowflake with ADLS as external stage, ingesting CSV, JSON, and Parquet data. Implements Bronze → Silver → Gold layers, CDC with Streams, data quality pipelines, and Tasks and Gold layer views for sales performance, customer segmentation, and product analytics.
Change Data Capture implementation (INSERT / UPDATE / DELETE) using Snowflake Streams with AWS S3 integration.
End-to-end automated data ingestion pipeline using Snowpipe — setup, configuration, and event-based triggering.
Querying and extracting nested JSON data in Snowflake using VARIANT data type and FLATTEN function.
Real-world SQL data cleaning — handling nulls, duplicates, standardization, and data type corrections.
Advanced SQL analytics on MLB player, team, and school data — window functions, aggregations, and performance insights.
Analyzing restaurant menu and order data to surface popular dishes, pricing trends, and customer spending patterns.
Real-world Airbnb dataset cleaned using Pandas — handling missing values, outliers, type conversions, and column normalization.
Amazon product data cleaned and preprocessed using Pandas — structured for downstream analytics or ML use.
Interactive HR dashboard covering employee headcount, attrition analysis, departmental performance, and workforce KPIs.
Visual analysis of personality survey data with dynamic slicers, trait distributions, and behavioral pattern breakdowns.
- 📧 Email: debashisdash1999@gmail.com
- 💼 LinkedIn: Choudhury Debashis Dash
✨ Always learning, always building — data tells the story, I make it clear.