GCP Data Engineer

Hartford, Connecticut

Inizio Partners
Apply for this Job
About the job GCP Data Engineer

Role - GCP Data Engineer (Visualization experience)

As a Data Engineer you will work on process of transforming raw data to a usable format, which is then further analyzed by other teams. Cleansing, organizing, and manipulating data using pipelines are some of the key responsibilities of a data engineer. You will also work on applying data engineering principles on the Google Cloud Platform to optimize its services. And create interactive dashboards/reports to present it to the stakeholders.

Role and Responsibilities:
  • Create and maintain optimal data pipeline architecture,
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and other data sources
  • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency, and other key business performance metrics.
  • Design intuitive and visually appearing visualizations to communicate complex data insights to users using Tableau, Streamlit, Dash Enterprise and Power BI
  • Use Flask to develop APIs that integrate data from various sources and facilitate automation of business processes.
  • Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
Candidate Profile:

Required Qualification and Skills:
  • Expert in DASH Enterprise in developing friendly UI and strong story telling experience.
  • Ability to conduct real-time data exploration, predictive modeling integration and a seamless deployment of machine learning modules to GCP.
  • Overall, 5-8 years of experience on ETL technologies
  • 3+ years of experience in data engineering technologies like SQL, Hadoop, Big Query, Dataproc, Composer
  • Strong Python/Pyspark Data Engineer Skills
  • Hands-on experience with data visualizations like Tableau, Power BI, Dash and Streamlit
  • Experience in building interactive dashboards, visualizations, and custom reports for business users.
  • Knowledge of Flask for developing APIs and automating data workflows
  • Experience in data automation and implemented workflows in a cloud environment.
  • Strong SQL ETL skills
  • GCP data engineer certified to be preferred.
  • Ability to understand and design the underlying data/schema
  • Strong communication skills to effectively communicate client updates.
Date Posted: 25 February 2025
Apply for this Job