Overall, 14 Years of experience in the fields of Big Data / BI and GCP  Certified Google Cloud Professional cloud Architect with 3 years of experience  Have Extensive Experience in IT data analytics projects, Hands on experience in migrating on premise ETLs to Google Cloud Platform (GCP) using cloud native tools such as Google BigQuery,DataFlow,DataFusion,Cloud Function,Pub\Sub, Composer,Airflow,Google Cloud Storage  Good experience in Data migration from on-prem MSSQL to Azure Cloud Snowflake DB  Completed Azure DP-900 certification  Knowledge in various ETL and Data Integration development tools like Informatica,Abi Intio and IBM Data Stage  Experience in Data Validation automation tool  Experience in Business Intelligence testing in various reports using Tableau, Power BI and Cognos framework tools  Good experience in Management tools as Azure Deveps, Jira,ALM and VSTS  Experience in preparing Test Strategy, Test Plan and Test estimation  Worked in Agile and Waterfall models  Good knowledge in good automation tools  Expertise in analyzing& reviewing business, functional and high-level technical requirements; designing detailed technical components for complex applications utilizing high-level architecture, design patterns and reusable code.  Gained expertise in design/architecture of Bigdata platforms and cloud technologies, building infrastructures with a secure solution for multi-site data centers and protecting/securing large volumes of data.  Good experience on multi-cluster architecture like on-premises to cloud, cloud to cloud architecture  Strong experience in delivering Big Data related projects using open-source technologies like Hadoop,py Spark, Sqoop, Hive, HBase, Kafka, Oozie, bigquery,gcs  Extensive Work Experience of infrastructure domains. E.g. Public Google Cloud Platform, Operating Systems like UNIX, Windows,  Extensive experience in implementing DevOps methodologies on Cloud platforms and through hands on experience in designing & creation of CI/CD pipelines with the tools like Jenkins, GIT, GitHub.  Having good Project Management skills which involves initiating, planning, executing, monitoring, controlling  Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.  Experience in optimizing ETL workflows.  Good experience on Hadoop tools related to Data warehousing like Hive and also involved in extracting the data from these tools on the cluster using pyspark JDBC API.  Skilled in executing programming code for intermediate to complex modules following development standards, planning and conducting code reviews for changes and enhancements that ensure standards compliance and systems interoperability.  Hands-on experience in working on Job Tracker, Task Tracker, Name Node, Data Node, Resource Manager, Node Manager, Application Master, YARN and MapReduce Concepts.  Excellence in managing Hive data warehouse tool-creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries.  Extensive expertise in Extracting and Loading data to various databases including Oracle, MS SQL Server, Teradata, Flat files, XML files using Talend.  Extensive expertise in developing XSD, XSLT and preparing XML files compatible to the xsd to parse the xml data into flat files to process into HDFS.  Good Experience in working with SerDe’s like Avro Format, Parquet format data.