Data Engineer
Connecting People with Their Purpose
At TalentBridge, we know success begins with personal connections, and our goal is to fully understand your values, motivations, and passions, then to connect you with companies whose values and motivations match your own.
Whether you're looking for temporary employment or a full-time career, the TalentBridge family is here to help.
About the Role
We're looking for a talented Data Engineer to join one of our Global 500 clients and be part of a growing team that plays a critical role in the organization.
In this role, your responsibilities will include:
Ingest, store, transform, and curate data from various internal and external sources including databases, flat files, and API endpoints
Query databases and data warehouses to access data using efficient SQL queries, PySpark, and ETL tools
Optimize solutions with a focus on usability, performance, flexibility, testability, and standardization
Demonstrate commitment to software development best practices, including documentation, reusable code design, test driven implementation, and integrated software testing
Document and demonstrate each solution by producing documentation, process flow diagrams, code comments, and clear code
Troubleshoot and analyze data inconsistencies, bugs, and performance related issues, often shifting priorities in a rapidly changing environment
Required Skills, Education, and Experience
Bachelor's degree required. Concentration in computer science, IT/IS, engineering, statistics, supply chain, business, or related field preferred but not required if experience in these fields can be substituted.
At least 2 years of experience in data engineering or related field (software engineering, data analytics, business intelligence, data science)
Ability to adapt quickly to an existing, complex environment
Ability to take initiative, handle multiple tasks, shift priorities, and meet deadlines
Ability to quickly learn new concepts and software as necessary, with little oversight
Strong interpersonal, written, and verbal communication skills
Technical proficiency in the following:
Strong SQL and Python skills required (including Pandas or PySpark)
Experience with ETL/ELT processes, data pipelines, orchestration, and workflow management required
Knowledge of relational database management systems and data warehousing concepts required
Experience with Git or knowledge of Azure DevOps, GitHub, or similar tools (e.g. Jira or GitLab) required
Experience with Databricks and/or Snowflake preferred
Experience with Azure ecosystem a plus - ADF, ADLS, Azure SQL DB
Experience with REST APIs a plus
Experience with Microsoft Excel, SAS, Oracle, and/or Microsoft SQL Server a plus
Specific Technical Skills:
SQL - Should be able to write SQL code very comfortably
DQL/DML statements, including a mastery of joins, conditionals/case when, aggregation, and filtering on datasets
DDL/defining objects like creating tables and views
Python
Should know the fundamentals, including the standard library and functions/OOP (i.e. should know what a class is, passing arguments into a method, etc)
Data engineering/data science experience preferred but also open to strong Python experience in web development or another related area
Need to be comfortable with lists, tuples, sets, and dictionaries
Databricks/PySpark and/or Pandas required
If you do not have experience with Databricks/PySpark, you must have experience with Pandas
You need to be familiar with the concept of a dataframe - getting data into it, manipulating it, running analysis on it, managing multiple dataframes together (joins, unions, etc), and so forth
Nice to Haves:
Familiarity with other libraries such as Requests (used commonly in this role for API calls)
Other advanced data science libraries such as NumPy and Matplotlib are always nice but not required for this role
Familiarity with advanced Data Structures
Date Posted: 20 May 2024
Job Expired - Click here to search for similar jobs