Join the Texas Health and Human Services Commission (HHSC) and be part of a team committed to creating a positive impact in the lives of fellow Texans. At HHSC, your contributions matter, and we support you at each stage of your life and work journey. Our comprehensive benefits package includes 100% paid employee health insurance for full-time eligible employees, a defined benefit pension plan, generous time off benefits, numerous opportunities for career advancement and more. Explore more details on the Benefits of Working at HHS webpage.
Functional Title: Data Engineer
Job Title: Data Architect I
Agency: Health & Human Services Comm
Department: Data and Technology
Posting Number: 3345
Closing Date: 08/15/2025
Posting Audience: Internal and External
Occupational Category: Computer and Mathematical
Salary Group: TEXAS-B-28
Salary Range: $7,716.66 - $10,000.00
Shift: Day
Additional Shift:
Telework: Eligible for Telework
Travel: Up to 5%
Regular/Temporary: Regular
Full Time/Part Time: Full time
FLSA Exempt/Non-Exempt: Exempt
Facility Location:
Job Location City: AUSTIN
Job Location Address: 4601 W GUADALUPE ST
Other Locations: Austin; Abilene; Amarillo; Arlington; Beaumont; Brenham; Bryan; Corpus Christi; Edinburg; El Paso; Fort Worth; Grand Prairie; Houston; Jacksonville; Lockhart; Lubbock; Marble Falls; Mexia; Nacogdoches; Port Arthur; San Angelo; San Antonio; Tyler; Waco; Wichita Falls
MOS Codes:
2659,8825,8846,14N,14NX,15AX,17S,181X,182X,1N0X1,1N1X1,1N4X1,255A,25B,26B,3D0X1,3D0X2,3D0X4,514A
5C0X1N,5C0X1S,5I0,62E,65FX,65WX,681X,682X,781X,782X,CT,CTT,CYB10,CYB11,ISM,IT
Brief Job Description:
The Texas Health and Human Services (HHS) Office of Inspector General (OIG) seeks a motivated and talented Data Engineer (Data Architect I) to join the Fraud Analytics unit within the OIG's Data Reviews Division.
The OIG's mission is to protect public resources and improve the health, safety, and well-being of Texans by preventing, detecting, auditing, inspecting, reviewing, and investigating fraud, waste, and abuse in the provision and delivery of all state health and human services and enforcing state law related to the provision of those services. The OIG's Data Reviews Division develops OIG initiatives through research and innovation, employing tools and processes that leverage the power of data to increase efficiency.
The Fraud Analytics unit plays a critical role in the OIG's mission by developing and implementing algorithms, artificial intelligence/machine learning (AI/ML) models, and advanced analytics to detect fraud, waste, and abuse, identify and respond to risks, and to generate insights and actionable leads that support investigations, audits, inspections, and reviews in state health and human services programs such as Medicaid, the Children's Health Insurance Program (CHIP), and Supplemental Nutrition Assistance Program (SNAP) and Temporary Assistance for Needy Families (TANF).
The Data Engineer will be responsible for designing, developing, integrating, maintaining, and optimizing scalable data pipelines that support these efforts and enable advanced analytics capabilities, ensuring that data analysts and data scientists have efficient, high-quality data to strengthen the OIG's data driven decision-making processes. This position reports to the Deputy Inspector General of Fraud Analytics and works under minimal supervision, with extensive latitude for the use of initiative and independent judgment.
Essential Job Functions:
Attends work on a regular and predictable schedule in accordance with agency leave policy and performs duties as assigned:
• Designs, builds, and manages efficient, scalable, and reliable data pipelines to support ETL/ELT processes and data integration efforts. Extracts data from multiple sources, and loads into a data warehouse, data store or other system and transforms raw data into structured, high-quality datasets that are readily usable by data analysts and data scientists to support analytics and data science use cases. Develops pipeline code using Python and SQL and sets standards for related metadata management. Implements and manages the data pipeline continuous integration and continuous delivery (CI/CD) processes using GitHub Enterprise. Manages data pipeline jobs throughout their lifecycle. Supports the task of deploying analytics and data science outputs into existing processes, applications, and dashboards. Develops and maintains a robust data architecture that supports the needs of data analysts and data scientists, while ensuring security and compliance standards are met. (40%)
• Develops and maintains documentation of best practices and standards for data engineering within the Office of Inspector General. Works closely with data analysts, data scientists, and other stakeholders to understand their data needs and ensure the accessibility, reliability, and scalability of data and to deliver necessary data infrastructure and resources. Implements processes and tools to monitor, validate, and ensure the accuracy and integrity of data used for analysis, with a focus on Databricks on Azure environments. Assists in the identification and integration of existing and potential new data sources, including HHS Snowflake data warehouses and Oracle databases. (30%)
• Performs Databricks administration tasks, including managing workspaces, clusters, and access controls. Optimizes Databricks environments for performance and cost-efficiency. Implements and maintains security protocols within the Databricks ecosystem. Manages integration between Databricks, Azure, Snowflake, and Oracle environments. (15%)
• Researches and evaluates emerging data engineering technologies and methodologies to enhance the organization's data capabilities, with a focus on Databricks, Azure, and open-source technologies, ensuring alignment with the OIG's strategic goals. Serves as a subject matter expert in health and human services and OIG program data. Serves as a liaison with external data engineering teams, the HHS Office of the Chief Data Architect, and technology vendors. Performs other duties necessary to achieve the mission of the Office of Inspector General. (15%) Knowledge, Skills, and Abilities (KSAs):
• Extensive knowledge of data engineering principles and practices, with a focus on cloud computing platforms, particularly Databricks, Azure, and Snowflake. Strong understanding of data modeling, data warehousing, data lake and lakehouse concepts, and querying relational databases. Proficiency in data wrangling, data storage, and data pipeline operations. Knowledge of data governance and quality standards, relational/non-relational data structures, theories, principles, and practices. Understanding of record keeping, including security procedures for handling, protecting, and distributing confidential data.
• Expert skill in Python for ETL/ELT processes and in SQL to query and optimize data queries to improve efficiency and reduce run times, especially in large, complex databases. Advanced experience with Databricks, including Delta Lake, Spark, and Databricks administration tasks. Proficiency in automating data collection and processing. Experience with Snowflake and Oracle data warehousing concepts and operations. Skill in Excel and PowerPoint.
• Skill in critical thinking, analyzing problems, and devising effective solutions.
• Verbal and written communication skills essential to effectively interact with data analysts, data scientists, OIG customers, and OIG leadership.
• Ability to design and implement complex data pipelines and architectures, particularly in Databricks and Azure environments.
• Ability to plan, organize, lead, manage and monitor data engineering projects from inception to completion.
• Ability to adapt to rapidly changing technologies and methodologies.
• Ability to effectively communicate complex data findings and analytical methodologies, both verbally and in writing, to technical and non-technical audiences.
• Ability to work effectively in teams that may cross departments and/or agencies. Ability to mentor and support team members in data engineering practices.
• Ability to work independently, exercise independent judgment, prioritize tasks, and manage multiple projects/assignments/responsibilities in a fast-paced environment under time constraints.
Registrations, Licensure Requirements or Certifications:
N/A
Initial Screening Criteria:
• Minimum of 3 years of experience in data engineering, including designing, building, and maintaining large-scale data pipelines and architectures using Python and SQL.
• Minimum of 3 years of experience with Databricks, including administration tasks.
• Experience with cloud platforms (Azure, AWS, GCP) preferred.
• Experience with VS Code preferred.
• Experience with Git for version control and collaborative development preferred.
• Experience with Snowflake and Oracle databases preferred.
• Experience in data engineering solutions in the healthcare domain preferred.
• Experience supporting AI/ML and advanced analytics/data science teams by designing and optimizing data pipelines for model training and inference preferred click apply for full job details
Date Posted: 22 April 2025
Apply for this Job