Must be a US Citizen & have a USA PassportSenior Data EngineerRemote WorkPrimary ResponsibilitiesDesign, develop, and maintain scalable data pipelines using Python, PySpark, and other modern programming languages to support both batch and streaming workloadsBuild and optimize data processing frameworks on cloud platforms such as Databricks or Snowflake, ensuring performance, reliability, and cost efficiencyDesign and implement robust data models, including transactional (OLTP) and dimensional (OLAP) schemas, to support analytics, reporting, and application integrationDevelop high quality SQL code including complex queries, stored procedures, and views, with a focus on performance tuning and efficient data access patternsCreate and manage workflow orchestration using Apache Airflow or similar tools, ensuring reliable scheduling, dependency management, and monitoringImplement and enforce data governance and metadata standards through tools such as Microsoft Purview, including data lineage, classification, cataloging, and security policiesBuild automated data quality and validation frameworks to ensure accuracy, completeness, and reliability of production datasetsCollaborate with cross functional teams including data architects, analysts, scientists, and business stakeholders to understand requirements and deliver scalable, well designed data solutionsLead technical design sessions and code reviews, promoting engineering best practices, reusability, and maintainabilitySupport cloud infrastructure and DevOps practices, including CI/CD pipelines, version control, testing automation, and environment managementMonitor and troubleshoot production data pipelines, proactively addressing issues, performance bottlenecks, and system failuresContribute to the evolution of the enterprise data platform, recommending tools, frameworks, and architectures to improve scalability and efficiencyYou'll be rewarded and recognized for your performance in an environment that will challenge you and give you clear direction on what it takes to succeed in your role as well as provide development for other roles you may be interested in.Required Qualifications7+ years of experience in data engineering, software engineering, or similar disciplinesHands‑on experience with Databricks or SnowflakeExperience with orchestration tools such as Apache AirflowExperience working with cloud ecosystems (Azure preferred; AWS/GCP acceptable)Advanced SQL skills and experience with OLTP and OLAP data modelingSolid understanding of modern data warehousing, data lake, and ELT/ETL design patternsFamiliarity with data governance tools, especially Microsoft PurviewSolid programming expertise in Python, PySpark, or similar languagesIf you are offered this position, you will be required to provide extensive personal information to obtain and maintain a suitability or determination of eligibility for a Confidential/Secret or Top Secret security clearance as a condition of your employmentUnited States CitizenshipPreferred QualificationsHealthcare industry experience, including claims, clinical, FHIR, HL7, or provider dataExperience with containerization (Docker, Kubernetes) for data workloadsExperience supporting machine learning workflows or analytical data science pipelinesKnowledge of distributed computing concepts and performance tuningThis is a remote position.#J-18808-Ljbffr