Post job

Data Engineer jobs at Danbury Hospital

- 439 jobs
  • Software Engineer

    R2 Global 4.3company rating

    Rochester, NY jobs

    A leading innovator in advanced laser technology is seeking a Software Engineer to join their growing Rochester engineering hub. This is a rare opportunity to work hands-on with cutting-edge scientific systems while building software that directly impacts real-world research, industrial applications, and next-generation engineered products. In this role, you'll collaborate closely with engineers, scientists, operations teams and business stakeholders to design, develop, and maintain a wide range of internal and product-level software tools. Your work will span intuitive user interfaces, automation solutions, data and KPI dashboards, and software that interacts with highly precise hardware systems. What you'll work on: • Building Python-based software for internal tools and product control systems • Creating user-facing interfaces that make complex technology simple to operate • Interfacing software with hardware, instruments, and electronic components • Designing dashboards and analytics to support operational decision-making • Contributing to automation, motion control, and system-integration projects • Maintaining high code standards through documentation and version control • Partnering with multi-disciplinary teams to define scope, requirements, and delivery timelines What makes this opportunity exciting: • Work in a highly respected, innovation-driven environment shaping the future of photonics • Engage with a diverse mix of engineering, R&D, and technology functions daily • See your work directly influence scientific and commercial applications • Join a collaborative team that values curiosity, creativity, and continuous learning • Enjoy the benefits of an on-site role within one of Rochester's strongest high-tech communities Ideal background: • Solid software development experience with strong Python skills • Experience building customer-facing or GUI-based software • Familiarity with hardware or instrument interfacing, embedded systems, or automation is a plus • Exposure to analytics, dashboards, or BI tools is helpful • Bachelor's degree in a technical field; advanced degree welcome but not required Compensation: Competitive salary up to $95,000, based on experience.
    $95k yearly 5d ago
  • Principal Software Engineer | Equity Trading | Portware 300k

    Living Talent 4.1company rating

    New York, NY jobs

    Hybrid: 4 days in office Compensation: 200k - 240k + 20-30% Bonus PTO: Unlimited Benefits: Exceptional Overview Implement automation strategies to improve trading workflows (Equities, Futures, FX, Options) New Development 30% Enhancements 50% Maintenance 20% Backend 70% Frontend 30% Build customizations and interfaces to Portware Interface with Traders and Portfolio Managers; collaborate with Product Owner Hands-on implementation (building out solutions, SDLC pipelines, code reviews, test case design) Provide architectural input - influence design aligning with future state Lead, mentor Tech Stack Java (50%) Java Swing (30% - for custom interfaces to EMS platform) SQL Server (10%) VB.net (10%) Portware EMS AWS, Snowflake Agile, Microservices TDD, Unit Testing (Pytest, UnitTest) GIT Co-pilot, Eclipse, VS Code, Jira, Confluence, Docker, Ideal experience: DevSecOps (ANT, Maven, Artifactory) Skills & Qualifications Trading workflow (Equities, Futures, Options, FX) Portware (or other EMS) 10+ years Fullstack Software Engineering 4+ years Architecture (microservices and other architectural concepts) Mastery of Java and Java frameworks Experience with Java Swing is a plus Cloud-native development expertise (ideally AWS) Skilled in troubleshooting and debugging performance issues Nice to haves: BS or MS in Computer Science, Math or related, CFA, CRD/CRIMS experience
    $120k-150k yearly est. 5d ago
  • Climate Data Consultant, Data and Analytics Section, DAPM, NYHQ, remote. req#585089

    Unicef 3.6company rating

    Remote

    If you are a committed, creative professional and are passionate about making a lasting difference for children, the world's leading children's rights organization would like to hear from you. For 70 years, UNICEF has been working on the ground in 190 countries and territories to promote children's survival, protection and development. The world's largest provider of vaccines fordeveloping countries, UNICEF supports child health and nutrition, good water and sanitation, quality basic education for all boys and girls, and the protection of children from violence, exploitation, and AIDS. UNICEF is funded entirely by the voluntary contributions of individuals, businesses, foundations and governments. UNICEF has over 12,000 staff in more than 145 countries. Consultancy: Climate Data Consultancy Duty Station: Data and Analytics Section, DAPM, NYHQ Duration: 01 Nov 2025 - 31 Oct 2026 Home/ Office Based: Remote BACKGROUND Purpose of Activity/ Assignment: UNICEF has established a new Global Child Hazard Database to estimate the exposure of children and critical infrastructures to single and multiple climate-related hazards. The purpose of this assignment is to enhance and scale the core data processing pipeline for the Database by integrating new data streams, improving processing efficiency, and preparing the system for broader deployment and increased data volume. Integrating different data sources and platforms in cloud-based server environment. Scope of Work: Scope of Work: Under the supervision and guidance of the Climate & Environment Data Science Specialist, the consultant will have the following duties and responsibilities: 1. Expand and Optimize Data Pipelines: • Scale the current data pipeline to create sub-national outputs with expanded pre-determined attributes, leveraging cloud resources using Google Earth Engine. • Enhance/expand the existing codebase for optimization and scaling, ensuring it can handle increased data volume and complexity. 2. Enhance Data Visualization: • Support the enhancement of the existing Google Earth Engine (GEE) data visualization application by developing and integrating server-side functions. • Support integrating different data platforms for report generation. 3. Data Analysis and Documentation: • Support downstream analysis and fulfill additional data processing requests. • Collaborate closely with the Statistics and Monitoring Manager and the Data Science Specialist to document and manage knowledge of all new and existing processes Terms of Reference / Key Deliverables: Work Assignment Overview/Deliverables and Outputs/Delivery deadline 1. Scaling the Core Pipeline for the Global Child Hazard Database (expanding the pipeline to include sub-national outputs.) - Pipeline Analysis and Architecture Blueprint (V0.9). Deliverable includes: (1) Comprehensive audit report of current pipeline performance, (2) Finalized list of 5+ sub-national data requirements, and (3) Formal architecture blueprint for GEE integration 30 Nov 2025 - GEE Processing Module Prototype (V1.0). Deliverable is a fully working Google Earth Engine script successfully ingesting and transforming one full-scale sub-national dataset, validating the core technology concept. 31 Dec 2025 - Production Module, Integration, and Documentation Package. Deliverable includes: (1) V2.0 module validated against 3 datasets (achieving >95% data quality), (2) Confirmed integration and API documentation for 3 internal platforms, including UNICEF's data warehouse, and (3) Final New Data Flow Manual. 31 Jan 2026 2. Support migrating to Google cloud - Enhanced GEE Application Release (V3.0). Deliverable includes: (1) tested server-side functions and (2) new interactive data visualization features integrated into the user interface 31 Mar 2026 3. Additional data to the Global Child Hazard Database - Database Enhancement and Ingestion Report. Deliverable is a report confirming the identification, validation, and verified successful ingestion of new, high-priority data sources into the Global Child Hazard Database 31 May 2026 4. Additional functions to visualization app - Automated Data Pipeline Implementation. Deliverable is the deployment and full documentation of at least two new, fully automated data pipelines for the sources identified in the Database Enhancement Report 31 Jul 2026 5. Optimization & Enhancement (Enhancing the existing codebase for efficiency and preparing for increased scale.) - Code Optimization and Scaling Report. Deliverable must demonstrate a 25% reduction in execution time for the top 5 most resource-intensive scripts, and document the implementation of at least two specific features designed for increased data volume. 31 Aug 2026 6. On-going country support, update of technical documentation & project finalization (Providing continuous support for analysis, ad-hoc requests, and knowledge transfer) - Project Finalization and Knowledge Transfer Package. Deliverable includes: (1) Twelve Monthly Support Logs, (2) Final Project Report, and (3) A minimum 1-hour recorded Knowledge Transfer Session with supporting materials. 31 Oct 2026 Qualifications Education: Bachelor's in computer science, data science, geospatial technology, or any other related discipline Experience working with Google Earth Engine Advanced proficiency in Python Knowledge/Expertise/Skills required *: A university degree in Computer Science, Data Science, Geospatial Technology, Remote Sensing, or any other closely related quantitative discipline is required. • Demonstrated 2+ years of professional expertise in geospatial analysis, data engineering, and developing production-ready data pipelines is required. • Advanced proficiency in Python and its geospatial packages, with a proven ability to write optimized, scalable code and independently debug complex programs is required. • Expert-level experience in Python-based Google Earth Engine (GEE) programming, including building and deploying complex GEE-based data processing and visualization applications is required. • Knowledge and experience deploying and managing data processing workloads on cloud-based computing platform (e.g. Google Cloud Platform, Azure) is required. • Knowledge and understanding of key issues and modeling challenges for climate and environmental data is an asset. • Strong interpersonal skills with internal and external stakeholders. • Excellent verbal and written communication skills. • Excellent data management skills (beyond basic Excel, including proficiency with data version control and database concepts). • Knowledge of climate, environment, and disaster risk reduction concepts and frameworks is an asset. • Familiarity with analytical frameworks and approaches to children and young people's rights, gender, and inclusion is an asset. • Fluent in spoken and written English. Additional official UN working languages are a plus. Requirements: Completed profile in UNICEF's e-Recruitment system and - Upload copy of academic credentials - Financial proposal that will include/ reflect : the costs per each deliverable and the total lump-sum for the whole assignment (in US$) to undertake the terms of reference. travel costs and daily subsistence allowance, if internationally recruited or travel is required as per TOR. Any other estimated costs: visa, health insurance, and living costs as applicable. Indicate your availability - Any emergent / unforeseen duty travel and related expenses will be covered by UNICEF. - At the time the contract is awarded, the selected candidate must have in place current health insurance coverage. - Payment of professional fees will be based on submission of agreed satisfactory deliverables. UNICEF reserves the right to withhold payment in case the deliverables submitted are not up to the required standard or in case of delays in submitting the deliverables on the part of the consultant. U.S. Visa information: With the exception of the US Citizens, G4 Visa and Green Card holders, should the selected candidate and his/her household members reside in the United States under a different visa, the consultant and his/her household members are required to change their visa status to G4, and the consultant's household members (spouse) will require an Employment Authorization Card (EAD) to be able to work, even if he/she was authorized to work under the visa held prior to switching to G4. Only shortlisted candidates will be contacted and advance to the next stage of the selection process For every Child, you demonstrate… UNICEF's core values of Commitment, Diversity and Integrity and core competencies in Communication, Working with People and Drive for Results. View our competency framework at: Here UNICEF offers reasonable accommodation for consultants/individual contractors with disabilities. This may include, for example, accessible software, travel assistance for missions or personal attendants. We encourage you to disclose your disability during your application in case you need reasonable accommodation during the selection process and afterwards in your assignment. UNICEF has a zero-tolerance policy on conduct that is incompatible with the aims and objectives of the United Nations and UNICEF, including sexual exploitation and abuse, sexual harassment, abuse of authority and discrimination. UNICEF also adheres to strict child safeguarding principles. All selected candidates will be expected to adhere to these standards and principles and will therefore undergo rigorous reference and background checks. Background checks will include the verification of academic credential(s) and employment history. Selected candidates may be required to provide additional information to conduct a background check. Remarks: Individuals engaged under a consultancy will not be considered “staff members” under the Staff Regulations and Rules of the United Nations and UNICEF's policies and procedures and will not be entitled to benefits provided therein (such as leave entitlements and medical insurance coverage). Their conditions of service will be governed by their contract and the General Conditions of Contracts for the Services of Consultants. Consultants are responsible for determining their tax liabilities and for the payment of any taxes and/or duties, in accordance with local or other applicable laws. The selected candidate is solely responsible to ensure that the visa (applicable) and health insurance required to perform the duties of the contract are valid for the entire period of the contract. Selected candidates are subject to confirmation of fully-vaccinated status against SARS-CoV-2 (Covid-19) with a World Health Organization (WHO)-endorsed vaccine, which must be met prior to taking up the assignment. It does not apply to consultants who will work remotely and are not expected to work on or visit UNICEF premises, programme delivery locations or directly interact with communities UNICEF works with, nor to travel to perform functions for UNICEF for the duration of their consultancy contracts.
    $70k-106k yearly est. 60d+ ago
  • MICS Data Harmonization Enhancement and Support for Tabulator Development Consultant, Data Collection Unit, Data and Analytics Section, DATA Team, DAPM, NYHQ, remote. Req#585091

    Unicef 3.6company rating

    Remote

    If you are a committed, creative professional and are passionate about making a lasting difference for children, the world's leading children's rights organization would like to hear from you. For 70 years, UNICEF has been working on the ground in 190 countries and territories to promote children's survival, protection and development. The world's largest provider of vaccines fordeveloping countries, UNICEF supports child health and nutrition, good water and sanitation, quality basic education for all boys and girls, and the protection of children from violence, exploitation, and AIDS. UNICEF is funded entirely by the voluntary contributions of individuals, businesses, foundations and governments. UNICEF has over 12,000 staff in more than 145 countries. Consultancy: MICS Data Harmonization Enhancement and Support for Tabulator Development Consultant Duty Station: Data Collection Unit, Data and Analytics Section, DATA Team; DAPM, NYHQ Duration: 15 Dec 2025 - 10 Nov 2026 Home/ Office Based: Remote BACKGROUND Purpose of Activity/ Assignment: UNICEF, as mandated by the United Nations General Assembly, is dedicated to advocating for the rights of every child, meeting their basic needs, and creating opportunities for their full development. A cornerstone of this mission is the Multiple Indicator Cluster Surveys (MICS) program - the largest source of internationally comparable data on children and women worldwide. MICS provides vital evidence for policymaking, program design, and progress monitoring toward global development commitments, including the Sustainable Development Goals (SDGs). Covering a wide range of thematic areas, MICS remains a key instrument for evidence-based decision-making at both national and international levels. While MICS data is publicly accessible, effective use of it often requires advanced statistical tools and expertise, which can limit its reach among policymakers and practitioners. To address this, UNICEF is developing the MICS Tabulator - an online platform designed to make MICS data easier to access, analyze, and visualize. The Tabulator will enable users to generate customized tabulations, pivot views, indicators, and visualizations directly online, without needing to download microdata or use specialized statistical software. By enhancing accessibility, the platform will empower policymakers, researchers, and development partners to leverage MICS data more effectively for informed action. The MICS Tabulator will play a pivotal role in broadening data use and dissemination. It will ensure that key insights on the well-being of children and women are accessible to a wider audience while maintaining international comparability and rigorous quality standards. In parallel, it will strengthen national capacity by enabling National Statistical Offices (NSOs) to conduct child-focused surveys with increasing autonomy, requiring only limited technical support. Ultimately, this initiative supports UNICEF's overarching goal of advancing evidence-based policymaking and improving outcomes for children and families worldwide. MICS surveys generate extensive datasets covering health, education, nutrition, child protection, and gender equality. Over time, adjustments to survey instruments and country-specific adaptations have led to structural variations across datasets, making cross-country and trend analyses more complex. To address these challenges, UNICEF has partnered with IPUMS at the University of Minnesota to harmonize MICS datasets across multiple rounds. This collaboration ensures consistency and comparability across countries and time, forming the backbone of the MICS Tabulator's harmonized database. Through this partnership, IPUMS has successfully harmonized 1,207 MICS datasets, utilizing translation tables, programming files, and SPSS codebooks to map variables into a unified framework. While highly effective, the current process relies heavily on external expertise. To ensure long-term sustainability and institutional capacity within UNICEF, a consultant will be engaged to review the existing harmonization workflows, documentation, and outputs. The consultant will help define a sustainable harmonization strategy for future MICS rounds - including standardized procedures, tools, and guidance for the MICS team - and support technical collaboration with the MICS Tabulator vendor by reviewing selected parts of the codebase and providing recommendations for improvement. Scope of Work: The consultant will build on the harmonization work already completed by IPUMS and support UNICEF in establishing a sustainable, in-house capacity to manage and extend data harmonization for future MICS survey rounds. IPUMS has developed a comprehensive library of translation tables and scripts used to convert raw MICS data into harmonized datasets. Maintaining and adapting these tools requires specialized knowledge. The consultant will review existing processes, document them clearly, and design a streamlined and future-proof approach for continued harmonization. The assignment requires a solid understanding of household survey methodologies, MICS-specific data processing workflows, and statistical programming tools. Working in close collaboration with UNICEF's technical teams, the consultant will ensure that the MICS Tabulator is built upon well-structured, standardized, and high-quality data - enhancing accessibility and usability for all stakeholders. Specific Roles and Responsibilities 1. Develop Templates and Guidance for Amendments Create templates and workflows for incorporating new variables or updates in translation tables. Produce detailed guidance for mapping new survey-specific variables to harmonized variable names and structures. Document illustrative examples of common amendments (e.g., new household characteristics, revised education categories). Provide training materials or recorded walkthroughs to facilitate internal capacity building. 2. Establish Processes and Tools for Future Harmonization Review IPUMS-produced code, translation tables, and harmonized outputs. Develop an efficient methodology and/or software scripts for processing new datasets and aligning them with the harmonized structure. Integrate automated validation and quality control checks into the harmonization workflow. Deliver comprehensive technical documentation and user manuals for UNICEF staff. Conduct training sessions or provide recorded materials to ensure sustainable knowledge transfer. 3. Support Vendor Code Review Provide technical support to the MICS Tabulator development vendor (Nagarro) by reviewing code related to dataset integration and harmonization. Identify and recommend improvements to enhance efficiency, maintainability, and alignment with UNICEF data standards. Collaborate with the vendor's development team to ensure smooth integration of harmonized datasets, contributing to selected code modules where necessary. 4. MICS Standard CAPI Listing and Mapping Application Develop a CSPro-based application for preparing MICS standard CAPI listings and mapping. Ensure the application automates key steps such as extracting data from survey inputs, mapping variables to the standardized structure, and performing basic validation checks to reduce manual work and errors Terms of Reference / Key Deliverables: Work Assignment Overview/Deliverables and Outputs/Delivery deadline 1. Amendment Templates, Guidance, and Training Materials - Standardized template with a detailed, step-by-step workflow for incorporating new variables or updating existing entries in translation tables for harmonized datasets including instructions, example entries, and a validation checklist. - Documented template with illustrative examples of workflows for adding new variables or updating translation tables for harmonized datasets. - Prepared training materials and/or recorded walkthroughs to facilitate internal capacity building of the MICS team 30 Apr 2026 2. Harmonization Process Package - Review report of IPUMS-produced code, translation tables, and harmonized outputs to ensure alignment with UNICEF standards. - Methodology documentation and 3-5 software scripts for processing new datasets and harmonizing them with the established structure. - Automated validation and quality control checks integrated into the harmonization workflow. - Documented technical guidance and user manuals for UNICEF staff. - At least 3 training sessions and/or recorded materials to ensure sustainable knowledge transfer. 30 Sept 2026 3. MICS Tabulator Vendor Code Review - Code review report for MICS Tabulator modules related to dataset integration and harmonization. - Written document with recommendations to enhance efficiency, maintainability, and alignment with UNICEF data standards. - Final report and supporting documentation of contributions and collaboration with the vendor's development team for the integration of harmonized datasets 10 Nov 2026 4. MICS Standard CAPI Listing and Mapping Application - Standard MICS CAPI listings and mapping application developed in CSPro 30 May 2026 Travel: One trip may be undertaken to meet with the MICS Tabulator vendor for coordination or progress review, if required and approved in advance by UNICEF Qualifications Education: Information Technologies, Statistics, Demography, or any other related technical field with expertise in data management Language Proficiency: Good communication skills in English Knowledge/Expertise/Skills required *: At least a Master's Degree or equivalent in Information Technologies, Statistics, Demography, or any other related technical field with expertise in data management. Minimum ten years' working experience in data processing in household surveys, preferably with prior MICS or DHS data processing experience. Expertise in programming with CSPro. Expertise in programming with SPSS and R. Strong IT and software development skills, including experience in reviewing and understanding vendor code for project development and integration. Excellent interpersonal skills Requirements: Completed profile in UNICEF's e-Recruitment system and - Upload copy of academic credentials - Financial proposal that will include/ reflect : the costs per each deliverable and the total lump-sum for the whole assignment (in US$) to undertake the terms of reference. travel costs and daily subsistence allowance, if internationally recruited or travel is required as per TOR. Any other estimated costs: visa, health insurance, and living costs as applicable. Indicate your availability - Any emergent / unforeseen duty travel and related expenses will be covered by UNICEF. - At the time the contract is awarded, the selected candidate must have in place current health insurance coverage. - Payment of professional fees will be based on submission of agreed satisfactory deliverables. UNICEF reserves the right to withhold payment in case the deliverables submitted are not up to the required standard or in case of delays in submitting the deliverables on the part of the consultant. U.S. Visa information: With the exception of the US Citizens, G4 Visa and Green Card holders, should the selected candidate and his/her household members reside in the United States under a different visa, the consultant and his/her household members are required to change their visa status to G4, and the consultant's household members (spouse) will require an Employment Authorization Card (EAD) to be able to work, even if he/she was authorized to work under the visa held prior to switching to G4. Only shortlisted candidates will be contacted and advance to the next stage of the selection process For every Child, you demonstrate… UNICEF's core values of Commitment, Diversity and Integrity and core competencies in Communication, Working with People and Drive for Results. View our competency framework at: Here UNICEF offers reasonable accommodation for consultants/individual contractors with disabilities. This may include, for example, accessible software, travel assistance for missions or personal attendants. We encourage you to disclose your disability during your application in case you need reasonable accommodation during the selection process and afterwards in your assignment. UNICEF has a zero-tolerance policy on conduct that is incompatible with the aims and objectives of the United Nations and UNICEF, including sexual exploitation and abuse, sexual harassment, abuse of authority and discrimination. UNICEF also adheres to strict child safeguarding principles. All selected candidates will be expected to adhere to these standards and principles and will therefore undergo rigorous reference and background checks. Background checks will include the verification of academic credential(s) and employment history. Selected candidates may be required to provide additional information to conduct a background check. Remarks: Individuals engaged under a consultancy will not be considered “staff members” under the Staff Regulations and Rules of the United Nations and UNICEF's policies and procedures and will not be entitled to benefits provided therein (such as leave entitlements and medical insurance coverage). Their conditions of service will be governed by their contract and the General Conditions of Contracts for the Services of Consultants. Consultants are responsible for determining their tax liabilities and for the payment of any taxes and/or duties, in accordance with local or other applicable laws. The selected candidate is solely responsible to ensure that the visa (applicable) and health insurance required to perform the duties of the contract are valid for the entire period of the contract. Selected candidates are subject to confirmation of fully-vaccinated status against SARS-CoV-2 (Covid-19) with a World Health Organization (WHO)-endorsed vaccine, which must be met prior to taking up the assignment. It does not apply to consultants who will work remotely and are not expected to work on or visit UNICEF premises, programme delivery locations or directly interact with communities UNICEF works with, nor to travel to perform functions for UNICEF for the duration of their consultancy contracts.
    $70k-106k yearly est. 60d+ ago
  • ETL Architect

    Unity Health Insurance 4.7company rating

    Wisconsin jobs

    Come Find Your Spark at Quartz! The ETL Architect will be responsible for the architecture, design, and implementation of data integration solutions and pipelines for the organization. This position will partner with multiple areas in the Enterprise Data Management team and the business to successfully translate business requirements into efficient and effective ETL implementations. This role will perform functional analysis, determining the appropriate data acquisition and ingestion methods, and design processes to populate various data platform layers. The ETL Architect will work with implementation stakeholders throughout the business to evaluate the state of data and constructs solutions that deliver data to enable analytics reporting capabilities in a reliable manner. Skills this position will utilize on a regular basis: * Informatica PowerCenter * Expert knowledge of SQL development * Python Benefits: * Opportunity to work with leading technology in the ever-changing, fast paced healthcare industry. * Opportunity to work across the organization interacting with business stakeholders. * Starting salary range based upon skills and experience: $107,500 - $134,400 - plus robust benefits package. Responsibilities * Architects, designs, enhances, and supports delivery of ETL solutions. * Architects and designs data acquisition, ingestion, transformation, and load solutions. * Identifies, develops, and documents ETL solution requirements to meet business needs. * Facilitates group discussions and joins solution design sessions with technical subject matter experts. * Develops, implements, and maintains standards and ETL design procedures. * Contributes to the design of the data models, data flows, transformation specifications, and processing schedules. * Coordinates ETL solution delivery and supports data analysis and information delivery staff in the design, development, and maintenance of data implementations. * Consults and provides direction on ETL architecture and the implementation of ETL solutions. * Queries, analyzes, and interprets complex data stored in the systems of record, enterprise data warehouse, and data marts. * Ensures work includes necessary audit, HIPAA compliance, and security controls. * Data Management * Collaborates with infrastructure and platform administrators to establish and maintain scalable and reliable data processing environment for the organization. * Identifies and triages data quality and performance issues from the ETL perspective and see them through to resolution. * Tests and validates components of the ETL solutions to ensure successful end-to-end delivery. * Participates in support rotation. Qualifications * Bachelor's degree with 8+ years of experience translating business requirements into business intelligence solutions, data visualization, and analytics solution design and development experience in a data warehouse and OLTP (Online Transaction Processing) environments, semantic layer modeling experience, and SQL programming experience. * OR associate degree with 11+ years of experience translating business requirements into business intelligence solutions, data visualization, and analytics solution design and development experience in a data warehouse and OLTP environments, semantic layer modeling experience, and SQL programming experience. * OR high school equivalence with 14+ years of experience translating business requirements into business intelligence solutions, data visualization, and analytics solution design and development experience in a data warehouse and OLTP environments, semantic layer modeling experience, and SQL programming experience. * Expert understanding of ETL concepts and commercially available enterprise data integration platforms (Informatica PowerCenter, Python) * Expert knowledge of SQL development * Expert knowledge of data warehousing concepts, design principles, associated data management and delivery requirements, and best practices * Expert problem solving and analytical skills * Ability to understand and communicate data management and integration concepts within IT and to the business and effectively interact with all internal and external parties including vendors and contractors * Ability to manage multiple projects simultaneously * Ability to work independently, under pressure, and be adaptable to change * Inquisitive and seek answers to questions without being asked Hardware and equipment will be provided by the company, but candidates must have access to high-speed, non-satellite Internet to successfully work from home. We offer an excellent benefit and compensation package, opportunity for career advancement and a professional culture built on the foundations of Respect, Responsibility, Resourcefulness and Relationships. To support a safe work environment, all employment offers are contingent upon successful completion of a pre-employment criminal background check. Quartz values and embraces diversity and is proud to be an Equal Employment Opportunity employer. All qualified applicants will receive consideration for employment without regard to race, religion, color, national origin, sex, gender identity or expression, sexual orientation, age, status as a protected veteran, among other things, or status as a qualified person with disability.
    $107.5k-134.4k yearly Auto-Apply 56d ago
  • Staff Data Engineer - Healthcare Data Infrastructure

    Qualified Disability Specialists 3.9company rating

    Remote

    Transform healthcare with us. At Qualified Health, we're redefining what's possible with Generative AI in healthcare. Our infrastructure provides the guardrails for safe AI governance, healthcare-specific agent creation, and real-time algorithm monitoring-working alongside leading health systems to drive real change. This is more than just a job. It's an opportunity to build the future of AI in healthcare, solve complex challenges, and make a lasting impact on patient care. If you're ambitious, innovative, and ready to move fast, we'd love to have you on board. Join us in shaping the future of healthcare. Job Summary: Qualified Health is seeking exceptional data engineers to join our growing team. You'll work at the intersection of healthcare data, cloud infrastructure, and AI-building robust, scalable data pipelines that power our platform. This role requires deep technical expertise in data engineering fundamentals, with opportunities to contribute across customer integrations, data quality and transformation, and platform automation depending on your strengths and interests. You'll collaborate with health systems and internal teams to deliver reliable healthcare data solutions using cutting-edge technologies. Key Responsibilities: Design, implement, and maintain production-grade data pipelines that process complex healthcare data at scale Build ETL/ELT solutions using modern cloud technologies and distributed computing frameworks Work with healthcare data standards including FHIR, HL7v2, and EHR-specific formats Ensure data quality, reliability, and compliance through comprehensive validation and monitoring Collaborate with cross-functional teams, customers, and stakeholders to deliver data solutions Debug and optimize pipeline performance in production environments Contribute to technical architecture decisions and platform evolution Document implementations and communicate technical concepts to varied audiences Required Qualifications: 6+ years of data engineering and/or related experience building and deploying production data pipelines Healthcare data experience working with clinical data standards (FHIR, HL7v2, CDA, X12) or EHR systems Strong SQL proficiency with ability to write complex queries, perform data modeling, and optimize performance Python expertise with modern development practices and frameworks Cloud platform experience (Azure, AWS, or GCP) with hands-on use of core data services ETL/ELT implementation using modern tools such as Databricks, Data Factory, Airflow, dbt, or similar platforms Proven problem-solving ability to independently debug complex data issues and drive solutions Excellent communication skills with ability to explain technical concepts clearly and document work thoroughly Bachelor's degree in Computer Science, Engineering, or related field Desirable Skills: Healthcare & Integration Expertise Direct experience establishing clinical data feeds from Epic, Cerner, eClinicalWorks, or other major EHR systems Knowledge of EHR-specific data formats and integration mechanisms (Epic Clarity, Chronicles, etc.) Experience with healthcare integration patterns including FHIR APIs, SFTP, Delta Sharing, Fabric External Shares, Snowflake Reader Accounts Customer-facing experience managing technical implementations and timelines Understanding of HIPAA and HITRUST compliance requirements in data infrastructure Data Quality & Transformation Healthcare data science or analytics background with understanding of clinical data models Experience designing sophisticated, configurable transformation logic and validation frameworks Expertise in data profiling, quality assessment, and building rule-based quality engines Strong instincts for identifying data quality issues and patterns in complex healthcare datasets Experience building data pipelines for AI/ML applications Platform Engineering & Automation Advanced experience with PySpark and distributed computing frameworks Infrastructure as Code (Terraform, CloudFormation, Pulumi, or similar) CI/CD pipeline development using GitHub Actions, Azure DevOps, GitLab, or similar Experience building reusable data engineering abstractions, frameworks, and developer tooling Production monitoring, alerting, and observability solutions (DataDog, Grafana, Prometheus, or similar) Track record establishing operational excellence standards for data systems Technical Depth Deep expertise in Azure cloud services (Databricks, Data Factory, Fabric, ADLS2, networking) Experience with modern data warehouse solutions (Snowflake, Databricks, Fabric) Real-time data processing and streaming architectures Advanced Python patterns for type safety and modern development Relevant certifications (Azure Data Engineer, Databricks, Snowflake, etc.) Master's degree in Computer Science, Engineering, or related field Strong candidates will excel in multiple areas above, demonstrating both breadth and depth in their expertise. Technical Environment: Our data infrastructure is built on modern cloud technologies including: Azure Databricks + Data Factory (plus Fabric and Snowflake integrations) PySpark for distributed data processing GitHub Actions + Terraform for CI/CD and Infrastructure as Code Python with type-safe patterns and modern frameworks Impact & Growth Opportunity: As a Senior Data Engineer at Qualified Health, you'll build the data foundation that powers our AI platform's ability to deliver insights to major health systems. You'll have the opportunity to shape our technical architecture and directly impact healthcare delivery, working with cutting-edge technologies in a rapidly growing company. This position offers significant visibility and growth potential as we scale our platform across the healthcare ecosystem. Why Join Qualified Health? This is an opportunity to join a fast-growing company and a world-class team, that is poised to change the healthcare industry. We are a passionate, mission-driven team that is building a category-defining product. We are backed by premier investors and are looking for founding team members who are excited to do the best work of their careers. Our employees are integral to achieving our goals so we are proud to offer competitive salaries with equity packages, robust medical/dental/vision insurance, flexible working hours, hybrid work options and an inclusive environment that fosters creativity and innovation. Our Commitment to Diversity Qualified Health is an equal opportunity employer. We believe that a diverse and inclusive workplace is essential to our success, and we are committed to building a team that reflects the world we live in. We encourage applications from all qualified individuals, regardless of race, color, religion, gender, sexual orientation, gender identity or expression, age, national origin, marital status, disability, or veteran status. Pay & Benefits: The pay range for this role is between $140,000 and $220,000, and will depend on your skills, qualifications, experience, and location. This role is also eligible for equity and benefits. Join our mission to revolutionize healthcare with AI. To apply, please send your resume through the application below.
    $140k-220k yearly Auto-Apply 7d ago
  • Forward Deployed Data Engineer

    Qventus 4.1company rating

    Remote

    On this journey for over 12 years, Qventus is leading the transformation of healthcare. We enable hospitals to focus on what matters most: patient care. Our innovative solutions harness the power of machine learning, generative AI, and behavioral science to deliver exceptional outcomes and empower care teams to anticipate and resolve issues before they arise. Our success in rapid scale across the globe is backed by some of the world's leading investors. At Qventus, you will have the opportunity to work with an exceptional, mission-driven team across the globe, and the ability to directly impact the lives of patients. We're inspired to work with healthcare leaders on our founding vision and unlock world-class medicine through world-class operations. #LI-JB1 The Role Forward Deployed Data Engineers at Qventus collaborate directly with clients to identify their most critical data challenges and design scalable, high-performance pipelines and architectures to solve them. Our customers depend on Qventus' data infrastructure for mission-critical healthcare operations, and projects often start with broad, high-impact questions like, “How can we unify real-time surgical, staffing, and patient flow data into a single source of truth?” or “What's the most efficient way to process and serve operational data for instant decision-making?” As a Data Engineer, you'll combine technical expertise in large-scale data systems with a deep understanding of operational needs to create solutions that bridge the gap between raw data and actionable insights. You'll work closely with data scientists, software engineers, and product teams to ensure our data pipelines are robust, efficient, and ready to support advanced analytics, AI models, and production-grade applications. You'll operate in small, agile teams with significant autonomy, taking projects from initial scoping and design through to deployment and ongoing optimization. A typical day might involve architecting cloud-based ETL workflows, optimizing query performance on multi-terabyte datasets, integrating disparate hospital data systems, or collaborating with client IT teams to ensure seamless adoption. Key Responsibilities Design, build, and maintain scalable data pipelines and architectures to support analytics, machine learning, and operational applications. Collaborate with cross-functional teams to translate complex operational needs into reliable, well-modeled datasets. Integrate and normalize data from multiple structured and unstructured healthcare sources (EHRs, scheduling systems, operational databases, etc.). Optimize query performance and data processing for speed, scalability, and cost efficiency. Implement best practices for data quality, governance, and security in compliance with healthcare regulations (e.g., HIPAA). Support deployment, monitoring, and troubleshooting of production data systems. What We're Looking For Proven experience as a data engineer or in a similar role, with a track record of building and maintaining large-scale data infrastructure. Strong proficiency in SQL and Python for data processing and pipeline development. Experience with cloud data platforms and services such as AWS (RDS, Redshift, Lambda, S3), GCP, or Azure. Knowledge of both relational and non-relational databases (PostgreSQL, MySQL, MongoDB, etc.). Familiarity with modern data workflow orchestration tools (Airflow, DBT, Dagster, etc.). Ability to work closely with both technical and non-technical stakeholders to gather requirements and deliver solutions. Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent experience. Bonus Points Experience working with healthcare data and integrating EHR, scheduling, or operational systems. Familiarity with real-time data processing frameworks (Kafka, Kinesis, Spark Streaming, Flink). Knowledge of data warehousing solutions like Snowflake or BigQuery. Hands-on experience with Databricks or similar data lakehouse platforms. Strong understanding of data privacy, compliance, and security in regulated environments. Experience mentoring peers or contributing to cross-functional technical strategy. Compensation for this role is based on market data and takes into account a variety of factors, including location, skills, qualifications, and prior relevant experience. Salary is just one part of the total rewards package at Qventus. We also offer a range of benefits and perks, including Open Paid Time Off, paid parental leave, professional development, wellness and technology stipends, a generous employee referral bonus, and employee stock option awards. Salary Range$140,000-$220,000 USD Qventus values diversity in its workforce and proudly upholds the principles of Equal Opportunity Employment . We welcome all qualified applicants and ensure fair consideration for employment without discrimination based on any legally protected characteristics, including, but not limited to: veteran status, uniformed service member status, race, color, religion, sex, sexual orientation, gender identity, age, pregnancy (including childbirth, lactation and related medical conditions), national origin or ancestry, citizenship or immigration status, physical or mental disability, genetic information (including testing and characteristics) or any other category protected by federal, state or local law (collectively, "protected characteristics"). Our commitment to equal opportunity employment applies to all persons involved in our operations and prohibits unlawful discrimination by any employee, including supervisors and co-workers. Qventus participates in the E-Verify program as required by law and is committed to providing reasonable accommodations to individuals with disabilities in compliance with Americans with Disabilities Act (ADA). In compliance with the California Consumer Privacy Act (CCPA), Qventus provides transparency into how applicant data is processed during the application process. Candidate information will be treated in accordance with our candidate privacy notice. *Benefits and perks are subject to plan documents and may change at the company's discretion. *Employment is contingent upon the satisfactory completion of our pre-employment background investigation and drug test.
    $140k-220k yearly Auto-Apply 7d ago
  • ETL Architect - Snowflake and SSIS Experience

    Northeast Georgia Health System 4.8company rating

    Remote

    Job Category: Administrative & Clerical Work Shift/Schedule: 8 Hr Morning - Afternoon Northeast Georgia Health System is rooted in a foundation of improving the health of our communities. ETL Architect is a high-level business intelligence professional that is a key part of the Data and Analytics Team and will be generally responsible for the dimensional model design, end to end ETL architecture design, ETL development, and implementation of data models and development of the Data Migration and Data Warehousing solutions. The ETL Architect is also responsible for requirements analysis, source data profiling, identifying data quality issues, designing transformation logic to address data quality issues, experienced with the concepts and/or implementation of metadata data capture, data lineage, master data management and reference data management. The ETL architect designs for data warehouse/data mart solutions should include appropriate staging databases to handle change data capture on the source data, error handling and audit balance control mechanisms. The ETL Architect will have extensive interaction with operational team members and leadership, and will be tasked with analysis of business data and analytics requirements, and the transformation of those requirements into analytics solutions. Working with customers and Management to determine priorities as well mentoring of other team members on ETL concepts and strategies is also an integral role of the position. Working with Infrastructure and DBA team members, ETL Architect is also responsible for keeping the ETL Environments updated periodically to the latest version of the ETL software. Minimum Job Qualifications Licensure or other certifications: Epic Revenue or Clinical Data Model or Cogito Project Manager Certificate required (must be obtained within 6 months of date of hire). Educational Requirements: Bachelors Degree in Computer Science or related field. In lieu of Bachelor's Degree related experience in the field, 11 years progressively responsible, professional job specific work experience and an equivalent combination of education and training that provides the required knowledge, skills and abilities, combined with demonstrated professional growth and achievement. Minimum Experience: A minimum of seven (7) years experience as a various BI and other analytic reporting software applications with strong preference for EPIC, SQL, and Crystal Report Writing. Prior experience providing production level support and/or development in business intelligence (BI)/enterprise data warehouse (EDW) environments. Other: Preferred Job Qualifications Preferred Licensure or other certifications: Preferred Educational Requirements: Preferred Experience: Experience in database development and management tasks such as creating indexes, functions, stored procedures, creating databases, file groups, users, configuring permissions and assigning roles to users. Working knowledge of data warehouse concepts, enterprise data management, modeling techniques and ability to create solutions, participate in strategy discussions and review them with customer as well as project stakeholders. Other: Job Specific and Unique Knowledge, Skills and Abilities Strong analytical and interpersonal skills; demonstrated ability to work with management and customers; proven ability to work independently or as a team member; strong problem solving and troubleshooting skills; initiative in enhancing system productivity, functionality, and performance. Advanced understanding of Data Migration/Data Warehousing/Data Mart delivery life cycle, from requirements definition to creating source to target mapping documents with transformation rules, Error handling design, data quality, design audit Balance control mechanisms to be able to automate jobs, notify proactively failures, write test cases and create templates to document all aspects of the projects throughout the lifecycle. Strong experience with concepts of data profiling, identifying and measuring data quality issues, advanced experience with Master data and reference data management techniques to eliminate hard coding within the ETL processes/jobs. Excellent communication (written, verbal and presentation) and people-facing skills, Proven ability to maintain positive peer, customer relationships, interacting positively and productively with teams across organizational lines. Demonstrates skills in leadership, critical thinking, and consensus and relationship building. Ability to review, re-engineer and/or tune legacy SQL queries and/or ETL jobs/processes to make them run efficiently and within a load window. Must be able to function in dynamic environment subject to changes in schedules and priorities, working on multiple projects/ issues concurrently. Responsibilities may require evening and weekend work in response to needs of the systems supported. Experience in estimation, specification gathering, review code/designs, general planning for execution, monitoring and control. Hands-on with ability to design, re-design (schemas, data models, entity relationships etc.), and probe and guide developers. Possess a solid understanding of data warehousing concepts, including relational and multidimensional modeling. Experience in designing, building, testing, executing and performance tuning of ETL and reporting applications, including automation of ETL jobs for recurring imports; fluent in Microsoft SQL. Experience in designing and developing database objects, tables, stored procedures, views, and triggers. Essential Tasks and Responsibilities Designs end to end ETL architecture for Data Migration, Data Warehousing or Data Mart Solutions using Microsoft SSIS tool. Ensure all ETL Jobs have adequate error handling and audit balance control mechanisms implemented to enable automation. Stands up a DW/ETL Center of Excellence to document and enforce best practices, standards, naming conventions etc. Creates a checklist to review all solutions developed to ensure adherence to the CoE standards. Mentors and trains staff to hold periodic CoE meetings. Implement a reference data management system using tools like Microsoft Master Data Management Services to eliminate hardcoding in ETL jobs. Maintains a service oriented approach with customers, providing follow up and responding to issues/ concerns in a timely manner. Answer staff and client questions in person, via email and via phone. Provide after-hours on-call support on a scheduled basis. Responsible for ensuring service level agreement adherence. Works with staff to adhere to department/organization policies and procedures, compliance and regulatory agency requirements. Works with leaders in IT as well as operations to identify data from other source systems to form richer data sets, including the capture of derived data to address gaps in source system repositories. Promotes and utilizes the data request and development process, including the development of an understanding of a data request, assessment of the appropriate tool to be used to provide the data, creation of a requirements document in partnership with the operational stakeholder and application analyst, development of the data extract and validation of the extract by the operational stakeholder to ensure quality, timeliness and accountability. Participates in and contributes to the data governance program - specifically, validating data lineage, capturing metadata for all existing and any newly developed assets (ETL Jobs), profile various data sources and address data quality issues, work with source system owners to reduce upstream data entry errors. Assists with all upgrade, update or enhancement efforts associated with the ETL tool. In partnership with the Manager of Business Intelligence and Analytics, serves as a mentor and advisor to the BI technical resources and other analysts and provides general project management skills to produce BI views that address customer's needs. Builds and maintains SQL scripts, indexes and complex queries for data analysis and extraction. Troubleshoots problems regarding the databases, applications, and development tools. Assists with the development and implementation of new data sources into the data warehouse or determine ways to be able to utilize data in other systems to support data and analytics for the organization. Conducts or coordinate tests to ensure that intelligence is consistent with defined needs. Responsible for data collection/extraction and developing methods for database design and validation reports. Prepares and maintains key documents and technical specifications for new and existing solutions. Documents specifications for business intelligence or information technology (IT) reports, dashboards, or other outputs. In collaboration with other team members and leadership, supports the evaluation, adoption and implementation of new methods, tools and technologies in order to meet current gaps in the organization-wide data and analytics infrastructure. Works closely with end-users to identify and explain data and reporting discrepancies. Prepares and maintains key documents and technical specifications for new and existing solutions. Provides solution design, perform reviews and participate strategic discussions. Provides application analysis and data modeling design to collect data or to extract data from various disparate systems. Monitors incident management queue. Efficiently troubleshoot, resolve, report and close active issues. Works on additional assignments as delegated by management. Physical Demands Weight Lifted: Up to 20 lbs, Occasionally 0-30% of time Weight Carried: Up to 20 lbs, Occasionally 0-30% of time Vision: Moderate, Frequently 31-65% of time Kneeling/Stooping/Bending: Occasionally 0-30% Standing/Walking: Frequently 31-65% Pushing/Pulling: Occasionally 0-30% Intensity of Work: Occasionally 0-30% Job Requires: Reading, Writing, Reasoning, Talking, Keyboarding, Driving Working at NGHS means being part of something special: a team invested in you as a person, an employee, and in helping you reach your goals. NGHS: Opportunities start here. Northeast Georgia Health System is an Equal Opportunity Employer and will not tolerate discrimination in employment on the basis of race, color, age, sex, sexual orientation, gender identity or expression, religion, disability, ethnicity, national origin, marital status, protected veteran status, genetic information, or any other legally protected classification or status.
    $91k-119k yearly est. Auto-Apply 7d ago
  • Staff Data Engineer

    Press Ganey Associates 4.7company rating

    Remote

    PG Forsta is the leading experience measurement, data analytics, and insights provider for complex industries-a status we earned over decades of deep partnership with clients to help them understand and meet the needs of their key stakeholders. Our earliest roots are in U.S. healthcare -perhaps the most complex of all industries. Today we serve clients around the globe in every industry to help them improve the Human Experiences at the heart of their business. We serve our clients through an unparalleled offering that combines technology, data, and expertise to enable them to pinpoint and prioritize opportunities, accelerate improvement efforts and build lifetime loyalty among their customers and employees. Like all great companies, our success is a function of our people and our culture. Our employees have world-class talent, a collaborative work ethic, and a passion for the work that have earned us trusted advisor status among the world's most recognized brands. As a member of the team, you will help us create value for our clients, you will make us better through your contribution to the work and your voice in the process. Ours is a path of learning and continuous improvement; team efforts chart the course for corporate success. Our Mission: We empower organizations to deliver the best experiences. With industry expertise and technology, we turn data into insights that drive innovation and action. Our Values: To put Human Experience at the heart of organizations so every person can be seen and understood. Energize the customer relationship: Our clients are our partners. We make their goals our own, working side by side to turn challenges into solutions. Success starts with me: Personal ownership fuels collective success. We each play our part and empower our teammates to do the same. Commit to learning: Every win is a springboard. Every hurdle is a lesson. We use each experience as an opportunity to grow. Dare to innovate: We challenge the status quo with creativity and innovation as our true north. Better together: We check our egos at the door. We work together, so we win together. We are seeking an experienced Staff Data Engineer to join our Unified Data Platform team. The ideal candidate will design, develop, and maintain enterprise-scale data infrastructure leveraging Azure and Databricks technologies. This role involves building robust data pipelines, optimizing data workflows, and ensuring data quality and governance across the platform. You will collaborate closely with analytics, data science, and business teams to enable data-driven decision-making. Duties & Responsibilities: Design, build, and optimize data pipelines and workflows in Azure and Databricks, including Data Lake and SQL Database integrations. Implement scalable ETL/ELT frameworks using Azure Data Factory, Databricks, and Spark. Optimize data structures and queries for performance, reliability, and cost efficiency. Drive data quality and governance initiatives, including metadata management and validation frameworks. Collaborate with cross-functional teams to define and implement data models aligned with business and analytical requirements. Maintain clear documentation and enforce engineering best practices for reproducibility and maintainability. Ensure adherence to security, compliance, and data privacy standards. Mentor junior engineers and contribute to establishing engineering best practices. Support CI/CD pipeline development for data workflows using GitLab or Azure DevOps. Partner with data consumers to publish curated datasets into reporting tools such as Power BI. Stay current with advancements in Azure, Databricks, Delta Lake, and data architecture trends. Technical Skills: Advanced proficiency in Azure 5+ years (Data Lake, ADF, SQL). Strong expertise in Databricks (5+ years), Apache Spark (5+ years), and Delta Lake (5+ years). Proficient in SQL (10+ years) and Python (5+ years); familiarity with Scala is a plus. Strong understanding of data modeling, data governance, and metadata management. Knowledge of source control (Git), CI/CD, and modern DevOps practices. Familiarity with Power BI visualization tool. Minimum Qualifications: Bachelor's or Master's degree in Computer Science, Data Science, or related field. 7+ years of experience in data engineering, with significant hands-on work in cloud-based data platforms (Azure). Experience building real-time data pipelines and streaming frameworks. Strong analytical and problem-solving skills. Proven ability to lead projects and mentor engineers. Excellent communication and collaboration skills. Preferred Qualifications: Master's degree in Computer Science, Engineering, or a related field. Exposure to machine learning integration within data engineering pipelines. Don't meet every single requirement? Studies have shown that women and people of color are less likely to apply to jobs unless they meet every single qualification. At PG Forsta we are dedicated to building a diverse, inclusive and authentic workplace, so if you're excited about this role but your past experience doesn't align perfectly with every qualification in the job description, we encourage you to apply anyways. You may be just the right candidate for this or other roles. Additional Information for US based jobs: Press Ganey Associates LLC is an Equal Employment Opportunity/Affirmative Action employer and well committed to a diverse workforce. We do not discriminate against any employee or applicant for employment because of race, color, sex, age, national origin, religion, sexual orientation, gender identity, veteran status, and basis of disability or any other federal, state, or local protected class. Pay Transparency Non-Discrimination Notice - Press Ganey will not discharge or in any other manner discriminate against employees or applicants because they have inquired about, discussed, or disclosed their own pay or the pay of another employee or applicant. However, employees who have access to the compensation information of other employees or applicants as a part of their essential job functions cannot disclose the pay of other employees or applicants to individuals who do not otherwise have access to compensation information, unless the disclosure is (a) in response to a formal complaint or charge, (b) in furtherance of an investigation, proceeding, hearing, or action, including an investigation conducted by the employer, or (c) consistent with the contractor's legal duty to furnish information. The expected base salary for this position ranges from $110,000 to $170,000. It is not typical for offers to be made at or near the top of the range. Salary offers are based on a wide range of factors including relevant skills, training, experience, education, and, where applicable, licensure or certifications obtained. Market and organizational factors are also considered. In addition to base salary and a competitive benefits package, successful candidates are eligible to receive a discretionary bonus or commission tied to achieved results. All your information will be kept confidential according to EEO guidelines. Our privacy policy can be found here: *****************************************
    $110k-170k yearly Auto-Apply 5d ago
  • Senior Data Engineer - Dallas, TX

    Photon Group 4.3company rating

    Remote

    We are seeking a highly motivated and experienced Data Engineer to play a key role in the design, development, and implementation of our next-generation data platform on Snowflake. You will be instrumental in building robust and scalable data pipelines, ensuring data quality, and enabling seamless data access for our Power BI reporting environment. Responsibilities: Design, develop, and maintain scalable and efficient ETL/ELT pipelines to ingest, transform, and load data into Snowflake from various source systems. Work closely with data analysts, business users, and other stakeholders to understand data requirements and translate them into technical solutions. Implement and enforce data quality standards and processes to ensure accurate and reliable data within the Snowflake environment. Optimize data models and query performance within Snowflake for efficient reporting and analysis. Develop and maintain data documentation, including data dictionaries and data flow diagrams. Contribute to the development of self-service data capabilities for business users. Stay up-to-date with the latest advancements in data engineering technologies and best practices, particularly within the Snowflake and cloud ecosystems. Troubleshoot and resolve data-related issues in a timely and efficient manner. Qualifications: Bachelor's degree in Computer Science, Engineering, or a related field. 10+ years of experience in data engineering or a similar role. Strong proficiency in SQL and experience working with relational databases (e.g., Oracle). Hands-on experience with cloud-based data warehousing platforms, Snowflake being highly preferred. Experience designing and implementing ETL/ELT processes using various tools and technologies. Experience with data quality frameworks and tools. Understanding of cloud computing concepts and platforms (e.g., AWS, Azure, GCP). Excellent problem-solving, analytical, and communication skills. Ability to work independently and collaboratively within a team environment. Experience with agile development methodologies is a plus. Compensation, Benefits and Duration Minimum Compensation: USD 38,000 Maximum Compensation: USD 134,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Sr Data Engineer - Tampa

    Photon Group 4.3company rating

    Remote

    Responsibilities Big Data - Spark, Hive, Java, CDP, 6 Years + with 3 Years of Dev experience on Big Data Analyze data requirements and identify disparate data sources required for consolidation and distribution. Document functional specifications and coordinate delivery of the same with technology team. Review logical and conceptual data models in alignment with business requirements. Work with the stakeholders to understand and gather requirements and produce business specifications Validate solution implementations and ensure they meet business and functional requirements. Provide production deployment support and investigate data quality issues. Work with various technology leads to ensure the gaps in the data completeness or accuracy are bridged. Qualifications Subject matter expertise in financial industry - wholesale loans /lending business OR Capital Markets or Finance or Risk Reporting Strong hands on experience with database and SQL is required. Excellent documentation, analytical skills to produce process flow diagrams, business modelling, and functional design. Proficiency in MS Office (Word, Excel, Visio, PowerPoint) with extensive experience using Excel for data analysis. Experience with Data tracing/ Lineage efforts Knowledge of logical and physical data model Compensation, Benefits and Duration Minimum Compensation: USD 42,000 Maximum Compensation: USD 148,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Data Engineer - Las Vegas, USA

    Photon Group 4.3company rating

    Remote

    Greetings Everyone Who are we? For the past 20 years, we have powered many Digital Experiences for the Fortune 500. Since 1999, we have grown from a few people to more than 4000 team members across the globe that are engaged in various Digital Modernization. For a brief 1 minute video about us, you can check ***************************** What will you do? What are we looking for? Job Description : Develop and maintain data pipelines, ELT processes, and workflow orchestration using Apache Airflow, Python and PySpark to ensure the efficient and reliable delivery of data. Design and implement custom connectors to facilitate the ingestion of diverse data sources into our platform, including structured and unstructured data from various document formats . Collaborate closely with cross-functional teams to gather requirements, understand data needs, and translate them into technical solutions. Implement DataOps principles and best practices to ensure robust data operations and efficient data delivery. Design and implement data CI/CD pipelines to enable automated and efficient data integration, transformation, and deployment processes. Monitor and troubleshoot data pipelines, proactively identifying and resolving issues related to data ingestion, transformation, and loading. Conduct data validation and testing to ensure the accuracy, consistency, and compliance of data. Stay up-to-date with emerging technologies and best practices in data engineering. Document data workflows, processes, and technical specifications to facilitate knowledge sharing and ensure data governance. Experience: Bachelor's degree in computer science, Engineering, or a related field 8 + years' experience in data engineering, ELT development, and data modeling. Proficiency in using Apache Airflow and Spark for data transformation, data integration, and data management. Experience implementing workflow orchestration using tools like Apache Airflow, SSIS or similar platforms. Demonstrated experience in developing custom connectors for data ingestion from various sources. Strong understanding of SQL and database concepts, with the ability to write efficient queries and optimize performance. Experience implementing DataOps principles and practices, including data CI/CD pipelines. Excellent problem-solving and troubleshooting skills, with a strong attention to detail. Effective communication and collaboration abilities, with a proven track record of working in cross-functional teams. Familiarity with data visualization tools Apache Superset and dashboard development. Understanding of distributed systems and working with large-scale datasets. Familiarity with data governance frameworks and practices. Knowledge of data streaming and real-time data processing technologies (e.g., Apache Kafka). Strong understanding of software development principles and practices, including version control (e.g., Git) and code review processes. Experience with Agile development methodologies and working in cross-functional Agile teams. Ability to adapt quickly to changing priorities and work effectively in a fast-paced environment. Excellent analytical and problem-solving skills, with a keen attention to detail. Strong written and verbal communication skills, with the ability to effectively communicate complex technical concepts to both technical and non-technical stakeholders. Required Skills - Python, Pyspark, Sql,Airflow, Trino, Hive, Snowflake, Agile Scrum Good to have- Linux,Openshift, Kubernentes, Superset Compensation, Benefits and Duration Minimum Compensation: USD 33,000 Maximum Compensation: USD 118,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 18d ago
  • Sr Data Engineer - New Jersey

    Photon Group 4.3company rating

    Remote

    We need a Senior Data Engineer 10+ years exp proficient in Spark, Scala/Java, and Hive, with extensive hands-on development experience in the Big Data Ecosystem. Key Responsibilities: Design, implement, and optimize highly performant data pipelines using Spark, Scala/Java, and Hive on platforms like Cloudera Data Platform (CDP) or other Hadoop echo systems. Take complete ownership of complex data engineering projects within the big data ecosystem, covering the entire lifecycle from initial design and development to deployment and ongoing maintenance. Develop robust and efficient Hive queries for extensive data analysis and reporting. Champion and enforce best practices and coding standards for new and existing data flows to ensure they are robust, scalable, secure, and maintainable using Spark, Scala/Java, and Hive within the big data ecosystem. Diagnose, troubleshoot, and resolve complex issues related to Spark, Scala/Java, and Hive applications and YARN resource management, implementing performance optimization solutions. Proactively collaborate with stakeholders, working closely to develop solutions with full commitment and accountability. Technical Skills & Experience: Proven hands-on development expertise with Apache Spark Strong programming proficiency in Scala and/or Java In-depth knowledge and practical experience with Hive, including query optimization and data analysis. Experience with data platforms such as Cloudera Data Platform (CDP) is highly desirable. Education: Bachelor's / Master's degree/University degree or equivalent experience
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Data Engineer- Smithfield, RI

    Photon Group 4.3company rating

    Remote

    We are seeking a skilled and passionate Data Engineer to join our team and play a vital role in building and maintaining our data infrastructure. The ideal candidate will have extensive experience with AWS cloud services, particularly EMR, and strong proficiency in Spark and PySpark for data processing and transformation. This role will focus on designing, developing, and optimizing data pipelines to support our growing data needs. Responsibilities: AWS Data Services: Design, implement, and manage data solutions on AWS, leveraging services such as EMR, S3, Glue, and others. Optimize AWS data infrastructure for performance, scalability, and cost-effectiveness. Implement best practices for data security and compliance on AWS. Apache Spark & PySpark: Develop and maintain scalable data pipelines using Apache Spark and PySpark. Perform data extraction, transformation, and loading (ETL/ELT) processes. Optimize Spark jobs for performance and efficiency. Develop and maintain data quality checks and validation processes. Amazon EMR: Configure and manage EMR clusters for large-scale data processing. Troubleshoot and resolve EMR cluster issues. Optimize EMR cluster configurations for performance and cost. Deploy and monitor spark applications on EMR. Data Pipeline Development: Design and implement robust and reliable data pipelines. Automate data ingestion, processing, and storage processes. Monitor data pipeline performance and troubleshoot issues. Work with various data sources, both structured and unstructured. Collaboration and Communication: Collaborate with data scientists, analysts, and other engineers to understand data requirements. Document data pipelines and infrastructure. Communicate effectively with technical and non-technical stakeholders. Participate in code reviews. Performance Optimization: Analyze query plans and optimize spark jobs. Monitor and tune data processing performance. Identify and resolve performance bottlenecks. Qualifications: Bachelor's degree in Computer Science, Data Science, or a related field (or equivalent experience). Minimum 6-9 years of experience in a Data Engineering role. Strong experience with Amazon Web Services (AWS) data services, particularly EMR. Proficiency in Apache Spark and PySpark for data processing. Experience with data warehousing and data lake concepts. Strong SQL skills. Experience with scripting languages (e.g., Python). Understanding of data modeling and database design principles. Experience with version control systems (e.g., Git). Strong problem-solving and troubleshooting skills. Excellent communication and collaboration skills. Experience with other big data technologies (e.g., Hadoop, Hive, Kafka) is a plus. Experience with data orchestration tools (ie airflow, step functions) is a plus. Compensation, Benefits and Duration Minimum Compensation: USD 37,000 Maximum Compensation: USD 130,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • SPARK Data Onboarding Engineer- Jersey City

    Photon Group 4.3company rating

    Remote

    Job Title: PySpark Data Engineer We are seeking a skilled PySpark Data Engineer to join our team and drive the development of robust data processing and transformation solutions within our data platform. You will be responsible for designing, implementing, and maintaining PySpark-based applications to handle complex data processing tasks, ensure data quality, and integrate with diverse data sources. The ideal candidate possesses strong PySpark development skills, experience with big data technologies, and the ability to work in a fast-paced, data-driven environment. Key Responsibilities: Data Engineering Development: Design, develop, and test PySpark-based applications to process, transform, and analyze large-scale datasets from various sources, including relational databases, NoSQL databases, batch files, and real-time data streams. Implement efficient data transformation and aggregation using PySpark and relevant big data frameworks. Develop robust error handling and exception management mechanisms to ensure data integrity and system resilience within Spark jobs. Optimize PySpark jobs for performance, including partitioning, caching, and tuning of Spark configurations. Data Analysis and Transformation: Collaborate with data analysts, data scientists, and data architects to understand data processing requirements and deliver high-quality data solutions. Analyze and interpret data structures, formats, and relationships to implement effective data transformations using PySpark. Work with distributed datasets in Spark, ensuring optimal performance for large-scale data processing and analytics. Data Integration and ETL: Design and implement ETL (Extract, Transform, Load) processes to ingest and integrate data from various sources, ensuring consistency, accuracy, and performance. Integrate PySpark applications with data sources such as SQL databases, NoSQL databases, data lakes, and streaming platforms Qualifications and Skills: Bachelor's degree in Computer Science, Information Technology, or a related field. 5+ years of hands-on experience in big data development, preferably with exposure to data-intensive applications. Strong understanding of data processing principles, techniques, and best practices in a big data environment. Proficiency in PySpark, Apache Spark, and related big data technologies for data processing, analysis, and integration. Experience with ETL development and data pipeline orchestration tools (e.g., Apache Airflow, Luigi). Strong analytical and problem-solving skills, with the ability to translate business requirements into technical solutions. Excellent communication and collaboration skills to work effectively with data analysts, data architects, and other team members. Compensation, Benefits and Duration Minimum Compensation: USD 43,000 Maximum Compensation: USD 151,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Data Engineer

    Photon Group 4.3company rating

    Remote

    Job Title: Data Engineer Design and develop microservices using Java Spring Boot for fund allocation, trade creation, portfolio rebalancing, and reporting services. Develop and expose RESTful APIs to handle cash movement, investor transactions, lending operations, and compliance alerts. Build a Kafka-based event-driven architecture, allowing asynchronous processing of key financial events (trade initiation, rebalancing signals, shareholder activity). Define and manage Kafka topics and consumer groups for modular and fault-tolerant data pipelines across trade and cash management domains. Conduct performance optimization for backend APIs to reduce latency in NAV (Net Asset Value) computation and high-frequency trade processing. Integrate APIs for Shareholder Activity, Position Lookup, Lending Services, and Warnings, enabling seamless multi-system orchestration. Perform in-depth analysis of legacy SQL stored procedures used in historical reporting and fund accounting. Extract undocumented business logic from complex SQL code and map legacy workflows to modern service-oriented architecture. Devise a migration and integration plan to transition critical stored procedures into microservices or event-driven components in the new system. Collaborate with data architects to restructure legacy database objects for compatibility with new cloud-native and real-time processing systems. Document data lineage, dependencies, and archival strategies to ensure business continuity and regulatory compliance. Build end-to-end UI automation test suites using Cypress and JavaScript to validate trading workflows and investor interactions. Create reusable test modules covering critical workflows including fund rebalancing, portfolio reporting, and user approvals. Integrate Cypress tests into CI/CD pipelines, supporting continuous testing and minimizing defects in production. Coordinate with QA and business stakeholders to translate requirements into test scenarios and acceptance criteria. Requirement: Applicants must have a Master's degree or foreign equivalent in Computer Science, Computer Applications, Information Technology, Engineering (any), or related fields. Travel/relocation required as jobs will be performed at various unanticipated locations throughout the United States.
    $95k-138k yearly est. Auto-Apply 34d ago
  • SPARK Data Reconciliation Engineer- Jersey City

    Photon Group 4.3company rating

    Remote

    Job Title: PySpark Data Reconciliation Engineer We're seeking a skilled PySpark Data Reconciliation Engineer to join our team and drive the development of robust data reconciliation solutions within our financial systems. You will be responsible for designing, implementing, and maintaining PySpark-based applications to perform complex data reconciliations, identify and resolve discrepancies, and automate data matching processes. The ideal candidate possesses strong PySpark development skills, experience with data reconciliation techniques, and the ability to integrate with diverse data sources and rules engines. Key Responsibilities: Data Reconciliation Development: Design, develop, and test PySpark-based applications to automate data reconciliation processes across various financial data sources, including relational databases, NoSQL databases, batch files, and real-time data streams. Implement efficient data transformation, matching algorithms (deterministic and heuristic) using PySpark and relevant big data frameworks. Develop robust error handling and exception management mechanisms to ensure data integrity and system resilience within Spark jobs. Data Analysis and Matching: Collaborate with business analysts and data architects to understand data requirements and matching criteria. Analyze and interpret data structures, formats, and relationships to implement effective data matching algorithms using PySpark. Work with distributed datasets in Spark, ensuring optimal performance for large-scale data reconciliation. Rules Engine Integration: Integrate PySpark applications with rules engines (e.g., Drools) or equivalent to implement and execute complex data matching rules. Develop PySpark code to interact with the rules engine, manage rule execution, and handle rule-based decision-making. Problem Solving and Gap Analysis: Collaborate with cross-functional teams to identify and analyze data gaps and inconsistencies between systems. Design and develop PySpark-based solutions to address data integration challenges and ensure data quality. Contribute to the development of data governance and quality frameworks within the organization. Qualifications and Skills: Bachelor's degree in Computer Science or a related field. 5+ years of hands-on experience in big data development, preferably with exposure to data-intensive applications. Strong understanding of data reconciliation principles, techniques, and best practices. Proficiency in PySpark, Apache Spark, and related big data technologies for data processing and integration. Experience with rules engine integration and development Strong analytical and problem-solving skills, with the ability to translate business requirements into technical solutions. Excellent communication and collaboration skills to work effectively with business analysts, data architects, and other team members. Familiarity with data streaming platforms (e.g., Kafka, Kinesis) and big data technologies (e.g., Hadoop, Hive, HBase) is a plus. Compensation, Benefits and Duration Minimum Compensation: USD 43,000 Maximum Compensation: USD 151,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Sr Data Engineer - Gen AI/ML - Tampa

    Photon Group 4.3company rating

    Remote

    Key Responsibilities: Model Development: Design and develop algorithms for generative models using deep learning techniques. Collaboration: Work with cross-functional teams to integrate generative AI solutions into existing systems. Research: Stay updated on the latest advancements in generative AI technologies and methodologies. Optimization: Fine-tune models for performance and efficiency. Troubleshooting: Address and resolve issues related to generative AI models and implementations. Documentation: Create and maintain comprehensive documentation for AI models and their applications. Communication: Explain complex technical concepts to non-technical stakeholders. Required Skills and Qualifications: 6+ years of strong background in machine learning and deep learning algorithms. Proficiency in programming languages such as Python, with experience in frameworks like TensorFlow and PyTorch. Familiarity with natural language processing (NLP) techniques and transformer models (e.g., GPT, BERT) Hands on experience with prompt structures and fine-tune model outputs to align with business needs and user expectations. Experience with generative AI techniques, including Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). Skills in data preprocessing and feature engineering for AI model training. Strong understanding of neural network architectures and optimization techniques. Experience in deploying AI models into production environments. Ability to stay updated with the latest advancements in generative AI research and incorporate them into work. Compensation, Benefits and Duration Minimum Compensation: USD 42,000 Maximum Compensation: USD 148,000 Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role. Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees. This position is not available for independent contractors No applications will be considered if received more than 120 days after the date of this post
    $95k-138k yearly est. Auto-Apply 60d+ ago
  • Senior Data Engineer

    Press Ganey Associates 4.7company rating

    Remote

    PG Forsta is the leading experience measurement, data analytics, and insights provider for complex industries-a status we earned over decades of deep partnership with clients to help them understand and meet the needs of their key stakeholders. Our earliest roots are in U.S. healthcare -perhaps the most complex of all industries. Today we serve clients around the globe in every industry to help them improve the Human Experiences at the heart of their business. We serve our clients through an unparalleled offering that combines technology, data, and expertise to enable them to pinpoint and prioritize opportunities, accelerate improvement efforts and build lifetime loyalty among their customers and employees. Like all great companies, our success is a function of our people and our culture. Our employees have world-class talent, a collaborative work ethic, and a passion for the work that have earned us trusted advisor status among the world's most recognized brands. As a member of the team, you will help us create value for our clients, you will make us better through your contribution to the work and your voice in the process. Ours is a path of learning and continuous improvement; team efforts chart the course for corporate success. Our Mission: We empower organizations to deliver the best experiences. With industry expertise and technology, we turn data into insights that drive innovation and action. Our Values: To put Human Experience at the heart of organizations so every person can be seen and understood. Energize the customer relationship: Our clients are our partners. We make their goals our own, working side by side to turn challenges into solutions. Success starts with me: Personal ownership fuels collective success. We each play our part and empower our teammates to do the same. Commit to learning: Every win is a springboard. Every hurdle is a lesson. We use each experience as an opportunity to grow. Dare to innovate: We challenge the status quo with creativity and innovation as our true north. Better together: We check our egos at the door. We work together, so we win together. Does analytical data flow through your veins? Are you fluent across cloud, on-premises, platforms, and technologies? Do you have deep healthcare experience? Do you love understanding strategic direction, creating design and direction, then getting fingers on the keyboard? This role is for a Senior Data Engineer with strong data architecture experience. This person will design, develop, and support solutions to transport, store, and analyze our analytical data. They will be key in evolving and maturing our enterprise data strategy and capabilities. Duties & Responsibilities: General Create data architecture, pipelines, and analytical solutions to meet software and data science requirements for various PG-MX Healthcare Products. Identify, evaluate, select, and prove out new technologies and toolsets. Create and execute Proofs of Concept and Proofs of Technology. Lead and direct the work of others in data dependent projects. Data Storage Design Collaborate with software development, business teams, analysts, and data scientists to establish data storage, pipeline, and structure requirements. Design, develop, and maintain ETL/ELT pipelines using Databricks (PySpark, Delta Lake, SQL). Implement Data Lakehouse architecture leveraging Databrick Unity Catalog, Delta Live Tables, and Workflows. Build and optimize data ingestion frameworks for structured and unstructured data from diverse sources. Identify and plan for data storage performance requirements. Optimize Databricks clusters, jobs, and queries for performance and cost efficiency. Application Interface and Data Storage Implementation Collaborate with software development, business teams, and data scientists to create and execute implementations. Identify impact of implementation on other applications and databases. Lead and mentor data engineers on data projects. Implement CI/CD for data pipelines using Git, Databricks Repos, and DevOps tools. Ensure Data quality, reliability, security compliances across environments. Master Data Management and Data Governance Build and evolve Trusted Record systems to manage entities across the enterprise. Design, implement, and evolve solutions around person identity management. Develop and enforce data governance, lineage, and cataloging standards. Mentors Data Engineers Identify areas of development and need. Provide targeted training and exploration for team members. Train and mentor data engineers on standards and best practices. Technical Skills: Hands on experience of Azure data technologies (Databricks, Data Factory, Stream Analytics, Data Lake Storage, Synapse), on-premises Microsoft tools (SQL DB and SSIS), and familiar with AWS data technologies. Proficiency in Python, SQL and distributed data processing frameworks (Spark), and familiar with C#, PowerShell, and APIs. Significant experience with analytical solutions in relational databases such as MS SQL Server, Oracle, and DB2 as well as experience with NoSQL databases and solutions such as data lakes, document-oriented databases, and graph databases. Strong understanding of data modeling, schema design, and ETL best practices Experience with data lineage, cataloging, and metadata management in Databricks and Unity Catalog. Skill in data modeling and experience with tools like ER/Studio or Erwin. Familiarity with version control (Git) and DevOps/CI-CD practices. Familiarity with SQL performance tuning and Spark optimization techniques. Excellent problem-solving and communication skills. Minimum Qualifications Minimum of 5 years Data Engineering experience in an enterprise environment. Bachelor's degree in technology or like field required. Don't meet every single requirement? Studies have shown that women and people of color are less likely to apply to jobs unless they meet every single qualification. At PG Forsta we are dedicated to building a diverse, inclusive and authentic workplace, so if you're excited about this role but your past experience doesn't align perfectly with every qualification in the job description, we encourage you to apply anyways. You may be just the right candidate for this or other roles. Additional Information for US based jobs: Press Ganey Associates LLC is an Equal Employment Opportunity/Affirmative Action employer and well committed to a diverse workforce. We do not discriminate against any employee or applicant for employment because of race, color, sex, age, national origin, religion, sexual orientation, gender identity, veteran status, and basis of disability or any other federal, state, or local protected class. Pay Transparency Non-Discrimination Notice - Press Ganey will not discharge or in any other manner discriminate against employees or applicants because they have inquired about, discussed, or disclosed their own pay or the pay of another employee or applicant. However, employees who have access to the compensation information of other employees or applicants as a part of their essential job functions cannot disclose the pay of other employees or applicants to individuals who do not otherwise have access to compensation information, unless the disclosure is (a) in response to a formal complaint or charge, (b) in furtherance of an investigation, proceeding, hearing, or action, including an investigation conducted by the employer, or (c) consistent with the contractor's legal duty to furnish information. The expected base salary for this position ranges from $100,000 to $120,000. It is not typical for offers to be made at or near the top of the range. Salary offers are based on a wide range of factors including relevant skills, training, experience, education, and, where applicable, licensure or certifications obtained. Market and organizational factors are also considered. In addition to base salary and a competitive benefits package, successful candidates are eligible to receive a discretionary bonus or commission tied to achieved results. All your information will be kept confidential according to EEO guidelines. Our privacy policy can be found here: *****************************************
    $100k-120k yearly Auto-Apply 53d ago
  • Principal Data Engineer - ML Platforms

    Altarum 4.5company rating

    Remote

    Altarum | Data & AI Center of Excellence (CoE) Altarum is building the future of data and AI infrastructure for public health - and we're looking for a Principal Data Engineer - ML Platforms to help lead the way. In this cornerstone role, you will design, build, and operationalize the modern data and ML platform capabilities that power analytics, evaluation, AI modeling, and interoperability across all Altarum divisions. If you want to architect impactful systems, enable data science at scale, and help ensure public health and Medicaid programs operate with secure, explainable, and trustworthy AI - this role is for you. What You'll Work On This role blends deep engineering with applied ML enablement: ML Platform Engineering: modern lakehouse architecture, pipelines, MLOps lifecycle Applied ML enablement: risk scoring, forecasting, Medicaid analytics NLP/Generative AI support: RAG, vectorization, health communications Causal ML operationalization: evaluation modeling workflows Responsible/Trusted AI engineering: model cards, fairness, compliance Your work ensures that Altarum's public health and Medicaid programs run on secure, scalable, reusable, and explainable data and AI infrastructure. What You'll Do Platform Architecture & Delivery Design and operate modern, cloud-agnostic lakehouse architecture using object storage, SQL/ELT engines, and dbt. Build CI/CD pipelines for data, dbt, and model delivery (GitHub Actions, GitLab, Azure DevOps). Implement MLOps systems: MLflow (or equivalent), feature stores, model registry, drift detection, automated testing. Engineer solutions in AWS and AWS GovCloud today, with portability to Azure Gov or GCP. Use Infrastructure-as-Code (Terraform, CloudFormation, Bicep) to automate secure deployments. Pipelines & Interoperability Build scalable ingestion and normalization pipelines for healthcare and public health datasets, including: FHIR R4 / US Core (strongly preferred) HL7 v2 (strongly preferred) Medicaid/Medicare claims & encounters (strongly preferred) SDOH & geospatial data (preferred) Survey, mixed-methods, and qualitative data Create reusable connectors, dbt packages, and data contracts for cross-division use. Publish clean, conformed, metrics-ready tables for Analytics Engineering and BI teams. Support Population Health in turning evaluation and statistical models into pipelines. Data Quality, Reliability & Cost Management Define SLOs and alerting; instrument lineage & metadata; ensure ≥95% of data tests pass. Perform performance and cost tuning (partitioning, storage tiers, autoscaling) with guardrails and dashboards. Applied ML Enablement Build production-grade pipelines for risk prediction, forecasting, cost/utilization models, and burden estimation. Develop ML-ready feature engineering workflows and support time-series/outbreak detection models. Integrate ML assets into standardized deployment workflows. Generative AI Enablement Build ingestion and vectorization pipelines for surveys, interviews, and unstructured text. Support RAG systems for synthesis, evaluation, and public health guidance. Enable Palladian Partners with secure, controlled-generation environments. Causal ML & Evaluation Engineering Translate R/Stata/SAS evaluation code into reusable pipelines. Build templates for causal inference workflows (DID, AIPW, CEM, synthetic controls). Support operationalization of ARA's applied research methods at scale. Responsible AI, Security & Compliance Implement Model Card Protocol (MCP) and fairness/explainability tooling (SHAP, LIME). Ensure compliance with HIPAA, 42 CFR Part 2, IRB/DUA constraints, and NIST AI RMF standards. Enforce privacy-by-design: tokenization, encryption, least-privilege IAM, and VPC isolation. Reuse, Shared-Services, and Enablement Develop runbooks, architecture diagrams, repo templates, and accelerator code. Pair with data scientists, analysts, and SMEs to build organizational capability. Provide technical guidance for proposals and client engagements. Your First 90 Days - You will make a meaningful impact fast. Expected outcomes include: Platform skeleton operational: repo templates, CI/CD, dbt project, MLflow registry, tests. Two pipelines in production (e.g., FHIR → analytics and claims normalization). One end-to-end CoE lighthouse MVP delivered (ingestion → model → metrics → BI). Completed playbooks for GovCloud deployment, identity/secrets, rollback, and cost control. Success Metrics (KPIs) Pipeline reliability meeting SLA/SLO targets. ≥95% data tests passing across pipelines. MVP dataset onboarding ≤ 4 weeks. Reuse of platform assets across ≥3 divisions. Cost optimization and budget adherence. What You'll Bring 7-10+ years in data engineering, ML platform engineering, or cloud data architecture. Expert in Python, SQL, dbt, and orchestration tools (Airflow, Glue, Step Functions). Deep experience with AWS + AWS GovCloud. CI/CD and IaC experience (Terraform, CloudFormation). Familiarity with MLOps tools (MLflow, Sagemaker, Azure ML, Vertex AI). Ability to operate in regulated environments (HIPAA, 42 CFR Part 2, IRB). Preferred: Experience with FHIR, HL7, Medicaid/Medicare claims, and/or SDOH datasets. Databricks, Snowflake, Redshift, Synapse. Event streaming (Kafka, Kinesis, Event Hubs). Feature store experience. Observability tooling (Grafana, Prometheus, OpenTelemetry). Experience optimizing BI datasets for Power BI. Logistical Requirements At this time, we will only accept candidates who are presently eligible to work in the United States and will not require sponsorship. Our organization requires that all work, for the duration of your employment, must be completed in the continental U.S. unless required by contract. If you're near one of our offices (Arlington, VA; Silver Spring, MD; or Novi, MI), you'll join us in person one day every other month (6 times per year) for a fun, purpose-driven Collaboration Day. These days are filled with creative energy, meaningful connection, and team brainstorming! Must be able to work during Eastern Time unless approved by your manager. Employees working remotely must have a dedicated, ergonomically appropriate workspace free from distractions with a mobile device that allows for productive and efficient conduct of business. Altarum is a nonprofit organization focused on improving the health of individuals with fewer financial resources and populations disenfranchised by the health care system. We work primarily on behalf of federal and state governments to design and implement solutions that achieve measurable results. We combine our expertise in public health and health care delivery with technology development and implementation, practice transformation, training and technical assistance, quality improvement, data analytics, and applied research and evaluation. Our innovative solutions and proven processes lead to better value and health for all. Altarum is an equal opportunity employer that provides employment and opportunities to all qualified employees and applicants without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, or any other characteristic protected by applicable law.
    $72k-98k yearly est. Auto-Apply 21d ago

Learn more about Danbury Hospital jobs

View all jobs