Pulastya Mitra

I'm

About

I have a master's degree and over 10+ years of experience in managing, transforming, and analyzing data using cloud platform such as Azure / AWS.


I am proficient in various data tools and technologies, such as SQL, Python, Spark, Azure, AWS and DBT.


GenAI enthusiast in generative AI models, including GPT, DALL-E, and Transformer-based architectures. Participated in IBM Azure Open AI Hackathon to build customer FAQ using Open AI embedding and storing it as Vector to build a chat bot.


I have successfully developed/worked on 25+ projects in various domains of data engineering, application, and data analysis.

Data Engineer & Data Analyst.

My expertise lies in developing data engineering solutions that modernize the legacy data processes. I can design and implement data pipelines that are efficient, scalable, and reliable.

  • Degree: Master
  • email: mitrapulastya@gmail.com
  • email: pulastyamitra@gmail.com
  • remote: Available

My expertise lies in developing data engineering solutions that modernize the legacy data processes. I can design and implement data pipelines that are efficient, scalable, and reliable.

Facts

Global exposure of working in the USA and managing onshore and offshore teams, developing client-centric materials, delivering training to client associates, leading the delivery of a large offshore team and product roll-out of customers.

Global Clients (Communication / Insurance / Industrial /Automobile)

Projects

Hours Of work experience

Global Exposure

Skills

I am proficient in various data tools and technologies, such as SQL, Python, Spark, Azure, AWS and DBT.

I have used visual exploration tools to create interactive dashboards and reports using matplotlib, seaborn, Bokeh, and Pandas. As well as Tableau, and Power BI.

ETL development (e.g., ingestion from APIs/databases and transformation)100%
Database design/modeling(MySQL, PostgreSQL, scylladb, MongoDB, MS SQL Server, Oracle) 90%
Working knowledge of distributed systems architecture (e.g., Spark, Event Hub) 80%
Statistical analysis methods (regression, clustering, etc.)70%
Programming languages:Python(Pandas + NumPy + Scikit-learn + Seaborn stack), .Net C#.70%
Cloud Platforms: Azure /AWS80%
Tableau70%
Data Build Tool(DBT)90%
Azure Data Factory (ADF)80%
Azure Databricks80%
Snowflake80%
PySpark80%
Delta Lake80%
Performance Improvements (SQL)95%
Power BI70%
Data Vault 2.090%

Resume

Global exposure by working in the USA and collaborating with onshore and offshore teams. I have also developed and led delivery of a large offshore team and rolled out products for customers.

Sumary

Pulastya Mitra

Seeking a Senior Data Engineer role to leverage my extensive experience in designing, developing, and implementing data solutions for complex challenges within a fast-paced environment.
I have a master's degree and over 10+ years of experience in managing, transforming, and analyzing data using cloud platforms such as Azure / AWS.
GenAI enthusiast in generative AI models, including GPT, DALL-E, and Transformer-based architectures. I participated in IBM Azure Open AI Hackathon to build customer FAQ using Open AI embedding and storing it as Vector to build a chat bot.

  • Kolkata, India
  • (+91) 9836629562
  • mitrapulastya@gmail.com

Education

Master of Computer Application (MCA)

2005 - 2007

Sikkim Manipal University, IN

MCA (Master of Computer Applications) is a professional master’s degree in computer science. It focuses on providing a theoretical as well as practical training to students in the related field.

Bachelor of Computer Application (BCA)

2002 - 2005

West Bengal University of Technology, IN

Bachelor of Computer Applications is an undergraduate course of 3 years that deals with various aspects of computer science and its applications in the real world. BCA subjects involve programming knowledge, developing applications, Analytical Thinking, and creativity.

Professional Experience

Data Engineering Manager| Cloud Technology | Gen AI

2025 - Present

πBy3, Pune, IN

  • Managing and guiding a young team responsible for migrating Legacy DataStage jobs to Databricks (using PySpark) with ADF orchestrating the process.
  • Developed an automated Python program to validate Azure Data Factory (ADF) pipeline and generate an Excel report for the testing team.
  • Overseeing an automated Python ingest process to transfer data from Salesforce, AWS RDS, and other sources to Snowflake. The process uses a configurable list of dimension and fact tables and supports both full load and incremental load options.
  • Guiding a team in organizing and developing proof-of-concept projects using Snowflake, Cortex AI, LLM, and Generative AI.
  • Overseeing and guiding a team to develop Gen AI proof-of-concept projects like chat-bot, early detection of Parkinson etc.

Senior Data Engineering - Advanced Analytics

2010 - 2025

IBM India Pvt. Ltd, Kolkata, IN

  • Worked as Lead Data Engineer to modernize existing DataMart to AWS, using AWS Glue ETL, CloudFormation, Redshift spectrum. The raw data was processed using Glue, DBT is used for transformation in Redshift.
  • My role involved leveraging the Data Build Tool (DBT) to streamline and automate data transformation processes, ensuring data integrity and efficiency. I collaborated with cross-functional teams to support data-driven decision-making, contributing to the organization's strategic objectives through robust data infrastructure management.
  • Managed a team responsible for converting T-SQL from an MIS SQL server into Data Build Tool (DBT), which involves a process of refactoring and migration. The initial step is to run the existing SQL transformations using DBT with minimal changes to establish a solid foundation. Following this, the business logic is segregated into stages or views on top of Databricks Lakehouse.
  • Worked on a project to migrate the existing Data Warehouse from Teradata to Snowflake. The process involved extracting data from Teradata, transforming it using DBT, and loading it into Snowflake.
  • Developed Cloud /On-Prem ETL pipelines and Data Architecture for batch and streaming systems from Claim systems to Claim Data warehouse.
  • Developed CAT (Catastrophic) Dashboard for real time claim registration using Databricks Delta Live tables. Showing FNOL Registration and payment details.
  • Worked and developed Claim Data Warehouse modernization base on Data Vault 2.0 principal using Azure Data Factory (ADF), Azure Databricks and Snowflake as Data Warehouse.
  • Developed a process to Ingest policy file from Azure Data Lake Storage into application database using Azure Data Factory (ADF).
  • Developed a process to Ingest Benefit / feature / coverage of each policy into application database using Azure Data Factory (ADF).

Data Engineering | Data Visualization |Application Developer

2008 - 2010

UshaComm India Pvt. Ltd, Kolkata, IN

  • Worked and developed TELECOM BI Analytics and decision management solutions mainly focusing on Revenue Analytics, Customer Analytics, and Sales & Marketing Analytics across OSS/BSS Revenue Chain and across the customer lifecycle for the TELECOM operators to achieve greater control on revenue, operation, and effectively differentiate themselves from competitors.
  • Decision support system (DSS) Measuring KPIs (ARPU, MoU, AMPU, Revenue, Churn, Customer Royalty, Ticket Closing Efficiency) by Customer Segment, Line of Business, Tariff Plans, Zone, and time (day, month, quarter, year) and displaying on Dashboard
  • Developed Reports on profiling, segmentation, pattern matching, trend, forecasting, and prediction (like identifying the most profitable customer/ product portfolio, Analyzing Churn using predictive modelling).
  • Worked and supported operational data to World Bank & Government of India Run Project RSBY.

Certification

  • All
  • Azure
  • Google
  • IBM
  • Linkedin
  • Others

Contact

Location:

Kolkata, IN

Call:

+91 9836629562

Resume:

Download