About the role
We are currently seeking a Senior Backend DataEngineer who can help drive the efforts to extend our web-scale data ingestion and processing pipeline. As part of this team, you will work as a tech leader and subject matter expert to improve the scalability, robustness, and value of our data ingestion and processing pipeline while leveraging the latest cloud technologies and AI / ML models.
What you'll do
Build our enterprise grade APIs: Our platform encompasses a broad spectrum of use-cases across a variety of domains. We need you to help build the APIs, backend services and data stores to empower our customers.
Ship commercial solutions: Our users have a huge job to do, at a massive scale. So, we need to build tremendous, powerful software that just works.
ML and big data in real life: Data and ML are at the core of what we do, and we are operationalizing them so that risk & intel, communications, and marketing teams at large companies and organizations can unlock their power. Doing so requires versatile, yet opinionated products - which require versatile, but opinionated engineers.
What you've done
Strong Industry Experience: Shown a steady upward career progression in the art of Software Engineering.
Complex Architectures: Design, implement, and maintain platforms, engines, or frameworks that others consume. Deliver APIs consumed by modern web applications with an understanding of the complexities and nuances of microservices and databases.
Distributed compute & big data: You've worked with distributed compute, and have tackled the performance challenges of large scale data. You've used serverless AWS components as building blocks like API Gateway, Lambda, Cognito, and DynamoDB.
Modern Backend Expertise: Build backend services and systems with languages like Python, Rust, or Go, and understand the complexities and nuances of CRUD, Graph API, and microservices.
CS Fundamentals: Earned at least a Bachelor's or Master's Degree in Computer Science, Computer Engineering, or equivalent.
Collaborate closely with product and sales teams to translate complex business needs into intuitive, performant tools.
Provide technical leadership and mentorship to a small but growing engineering team, setting coding standards, guiding architectural decisions, and fostering a culture of collaboration and innovation.
What we offer
Grow as a professional by working with emerging technologies in a high-demand industry segment
Drive company direction at a company that encourages input from all levels to help shape our trajectory
Experience working at a technology company in which the mission focus is on tackling one of society's greatest challenges
Learn from industry and government experts who value initiative and encourage innovation
Enjoy the flexibility to work remotely, unlimited PTO, company (and two personal-choice) holidays
Receive excellent benefits, including 100% employer-paid employee health premiums (50% for dependents), 401(k), and FSA
The Senior Backend Engineer / DataEngineer position offers a base salary range of $150,000 to $240,000, along with equity and comprehensive benefits. Please note that this range serves as a guideline; actual total compensation may vary based on factors such as experience, skill set, qualifications, and geographic location.
$150k-240k yearly 23d ago
Looking for a job?
Let Zippia find it for you.
Forward Deployed Data Engineer
Qventus 4.1
Remote
On this journey for over 12 years, Qventus is leading the transformation of healthcare. We enable hospitals to focus on what matters most: patient care. Our innovative solutions harness the power of machine learning, generative AI, and behavioral science to deliver exceptional outcomes and empower care teams to anticipate and resolve issues before they arise.
Our success in rapid scale across the globe is backed by some of the world's leading investors. At Qventus, you will have the opportunity to work with an exceptional, mission-driven team across the globe, and the ability to directly impact the lives of patients. We're inspired to work with healthcare leaders on our founding vision and unlock world-class medicine through world-class operations. #LI-JB1
The Role
Forward Deployed DataEngineers at Qventus collaborate directly with clients to identify their most critical data challenges and design scalable, high-performance pipelines and architectures to solve them. Our customers depend on Qventus' data infrastructure for mission-critical healthcare operations, and projects often start with broad, high-impact questions like, “How can we unify real-time surgical, staffing, and patient flow data into a single source of truth?” or “What's the most efficient way to process and serve operational data for instant decision-making?”
As a DataEngineer, you'll combine technical expertise in large-scale data systems with a deep understanding of operational needs to create solutions that bridge the gap between raw data and actionable insights. You'll work closely with data scientists, software engineers, and product teams to ensure our data pipelines are robust, efficient, and ready to support advanced analytics, AI models, and production-grade applications.
You'll operate in small, agile teams with significant autonomy, taking projects from initial scoping and design through to deployment and ongoing optimization. A typical day might involve architecting cloud-based ETL workflows, optimizing query performance on multi-terabyte datasets, integrating disparate hospital data systems, or collaborating with client IT teams to ensure seamless adoption.
Key Responsibilities
Design, build, and maintain scalable data pipelines and architectures to support analytics, machine learning, and operational applications.
Collaborate with cross-functional teams to translate complex operational needs into reliable, well-modeled datasets.
Integrate and normalize data from multiple structured and unstructured healthcare sources (EHRs, scheduling systems, operational databases, etc.).
Optimize query performance and data processing for speed, scalability, and cost efficiency.
Implement best practices for data quality, governance, and security in compliance with healthcare regulations (e.g., HIPAA).
Support deployment, monitoring, and troubleshooting of production data systems.
What We're Looking For
Proven experience as a dataengineer or in a similar role, with a track record of building and maintaining large-scale data infrastructure.
Strong proficiency in SQL and Python for data processing and pipeline development.
Experience with cloud data platforms and services such as AWS (RDS, Redshift, Lambda, S3), GCP, or Azure.
Knowledge of both relational and non-relational databases (PostgreSQL, MySQL, MongoDB, etc.).
Familiarity with modern data workflow orchestration tools (Airflow, DBT, Dagster, etc.).
Ability to work closely with both technical and non-technical stakeholders to gather requirements and deliver solutions.
Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent experience.
Bonus Points
Experience working with healthcare data and integrating EHR, scheduling, or operational systems.
Familiarity with real-time data processing frameworks (Kafka, Kinesis, Spark Streaming, Flink).
Knowledge of data warehousing solutions like Snowflake or BigQuery.
Hands-on experience with Databricks or similar data lakehouse platforms.
Strong understanding of data privacy, compliance, and security in regulated environments.
Experience mentoring peers or contributing to cross-functional technical strategy.
Compensation for this role is based on market data and takes into account a variety of factors, including location, skills, qualifications, and prior relevant experience. Salary is just one part of the total rewards package at Qventus. We also offer a range of benefits and perks, including Open Paid Time Off, paid parental leave, professional development, wellness and technology stipends, a generous employee referral bonus, and employee stock option awards.
Salary Range$140,000-$220,000 USD
Qventus values diversity in its workforce and proudly upholds the principles of Equal Opportunity Employment . We welcome all qualified applicants and ensure fair consideration for employment without discrimination based on any legally protected characteristics, including, but not limited to: veteran status, uniformed service member status, race, color, religion, sex, sexual orientation, gender identity, age, pregnancy (including childbirth, lactation and related medical conditions), national origin or ancestry, citizenship or immigration status, physical or mental disability, genetic information (including testing and characteristics) or any other category protected by federal, state or local law (collectively, "protected characteristics"). Our commitment to equal opportunity employment applies to all persons involved in our operations and prohibits unlawful discrimination by any employee, including supervisors and co-workers.
Qventus participates in the E-Verify program as required by law and is committed to providing reasonable accommodations to individuals with disabilities in compliance with Americans with Disabilities Act (ADA). In compliance with the California Consumer Privacy Act (CCPA), Qventus provides transparency into how applicant data is processed during the application process. Candidate information will be treated in accordance with our candidate privacy notice.
*Benefits and perks are subject to plan documents and may change at the company's discretion.
*Employment is contingent upon the satisfactory completion of our pre-employment background investigation and drug test.
$140k-220k yearly Auto-Apply 46d ago
ETL Architect - Snowflake and SSIS Experience
Northeast Georgia Health System 4.8
Remote
Job Category:
Administrative & Clerical
Work Shift/Schedule:
8 Hr Morning - Afternoon
Northeast Georgia Health System is rooted in a foundation of improving the health of our communities.
ETL Architect is a high-level business intelligence professional that is a key part of the Data and Analytics Team and will be generally responsible for the dimensional model design, end to end ETL architecture design, ETL development, and implementation of data models and development of the Data Migration and Data Warehousing solutions. The ETL Architect is also responsible for requirements analysis, source data profiling, identifying data quality issues, designing transformation logic to address data quality issues, experienced with the concepts and/or implementation of metadata data capture, data lineage, master data management and reference data management. The ETL architect designs for data warehouse/data mart solutions should include appropriate staging databases to handle change data capture on the source data, error handling and audit balance control mechanisms. The ETL Architect will have extensive interaction with operational team members and leadership, and will be tasked with analysis of business data and analytics requirements, and the transformation of those requirements into analytics solutions. Working with customers and Management to determine priorities as well mentoring of other team members on ETL concepts and strategies is also an integral role of the position. Working with Infrastructure and DBA team members, ETL Architect is also responsible for keeping the ETL Environments updated periodically to the latest version of the ETL software.
Minimum Job Qualifications
Licensure or other certifications: Epic Revenue or Clinical Data Model or Cogito Project Manager Certificate required (must be obtained within 6 months of date of hire).
Educational Requirements: Bachelors Degree in Computer Science or related field. In lieu of Bachelor's Degree related experience in the field, 11 years progressively responsible, professional job specific work experience and an equivalent combination of education and training that provides the required knowledge, skills and abilities, combined with demonstrated professional growth and achievement.
Minimum Experience: A minimum of seven (7) years experience as a various BI and other analytic reporting software applications with strong preference for EPIC, SQL, and Crystal Report Writing. Prior experience providing production level support and/or development in business intelligence (BI)/enterprise data warehouse (EDW) environments.
Other:
Preferred Job Qualifications
Preferred Licensure or other certifications:
Preferred Educational Requirements:
Preferred Experience: Experience in database development and management tasks such as creating indexes, functions, stored procedures, creating databases, file groups, users, configuring permissions and assigning roles to users. Working knowledge of data warehouse concepts, enterprise data management, modeling techniques and ability to create solutions, participate in strategy discussions and review them with customer as well as project stakeholders.
Other:
Job Specific and Unique Knowledge, Skills and Abilities
Strong analytical and interpersonal skills; demonstrated ability to work with management and customers; proven ability to work independently or as a team member; strong problem solving and troubleshooting skills; initiative in enhancing system productivity, functionality, and performance.
Advanced understanding of Data Migration/Data Warehousing/Data Mart delivery life cycle, from requirements definition to creating source to target mapping documents with transformation rules, Error handling design, data quality, design audit Balance control mechanisms to be able to automate jobs, notify proactively failures, write test cases and create templates to document all aspects of the projects throughout the lifecycle.
Strong experience with concepts of data profiling, identifying and measuring data quality issues, advanced experience with Master data and reference data management techniques to eliminate hard coding within the ETL processes/jobs.
Excellent communication (written, verbal and presentation) and people-facing skills, Proven ability to maintain positive peer, customer relationships, interacting positively and productively with teams across organizational lines. Demonstrates skills in leadership, critical thinking, and consensus and relationship building.
Ability to review, re-engineer and/or tune legacy SQL queries and/or ETL jobs/processes to make them run efficiently and within a load window.
Must be able to function in dynamic environment subject to changes in schedules and priorities, working on multiple projects/ issues concurrently. Responsibilities may require evening and weekend work in response to needs of the systems supported.
Experience in estimation, specification gathering, review code/designs, general planning for execution, monitoring and control.
Hands-on with ability to design, re-design (schemas, data models, entity relationships etc.), and probe and guide developers.
Possess a solid understanding of data warehousing concepts, including relational and multidimensional modeling.
Experience in designing, building, testing, executing and performance tuning of ETL and reporting applications, including automation of ETL jobs for recurring imports; fluent in Microsoft SQL.
Experience in designing and developing database objects, tables, stored procedures, views, and triggers.
Essential Tasks and Responsibilities
Designs end to end ETL architecture for Data Migration, Data Warehousing or Data Mart Solutions using Microsoft SSIS tool. Ensure all ETL Jobs have adequate error handling and audit balance control mechanisms implemented to enable automation.
Stands up a DW/ETL Center of Excellence to document and enforce best practices, standards, naming conventions etc. Creates a checklist to review all solutions developed to ensure adherence to the CoE standards. Mentors and trains staff to hold periodic CoE meetings.
Implement a reference data management system using tools like Microsoft Master Data Management Services to eliminate hardcoding in ETL jobs.
Maintains a service oriented approach with customers, providing follow up and responding to issues/ concerns in a timely manner. Answer staff and client questions in person, via email and via phone. Provide after-hours on-call support on a scheduled basis. Responsible for ensuring service level agreement adherence.
Works with staff to adhere to department/organization policies and procedures, compliance and regulatory agency requirements.
Works with leaders in IT as well as operations to identify data from other source systems to form richer data sets, including the capture of derived data to address gaps in source system repositories.
Promotes and utilizes the data request and development process, including the development of an understanding of a data request, assessment of the appropriate tool to be used to provide the data, creation of a requirements document in partnership with the operational stakeholder and application analyst, development of the data extract and validation of the extract by the operational stakeholder to ensure quality, timeliness and accountability.
Participates in and contributes to the data governance program - specifically, validating data lineage, capturing metadata for all existing and any newly developed assets (ETL Jobs), profile various data sources and address data quality issues, work with source system owners to reduce upstream data entry errors.
Assists with all upgrade, update or enhancement efforts associated with the ETL tool.
In partnership with the Manager of Business Intelligence and Analytics, serves as a mentor and advisor to the BI technical resources and other analysts and provides general project management skills to produce BI views that address customer's needs.
Builds and maintains SQL scripts, indexes and complex queries for data analysis and extraction.
Troubleshoots problems regarding the databases, applications, and development tools.
Assists with the development and implementation of new data sources into the data warehouse or determine ways to be able to utilize data in other systems to support data and analytics for the organization.
Conducts or coordinate tests to ensure that intelligence is consistent with defined needs.
Responsible for data collection/extraction and developing methods for database design and validation reports.
Prepares and maintains key documents and technical specifications for new and existing solutions.
Documents specifications for business intelligence or information technology (IT) reports, dashboards, or other outputs.
In collaboration with other team members and leadership, supports the evaluation, adoption and implementation of new methods, tools and technologies in order to meet current gaps in the organization-wide data and analytics infrastructure.
Works closely with end-users to identify and explain data and reporting discrepancies.
Prepares and maintains key documents and technical specifications for new and existing solutions.
Provides solution design, perform reviews and participate strategic discussions.
Provides application analysis and data modeling design to collect data or to extract data from various disparate systems.
Monitors incident management queue. Efficiently troubleshoot, resolve, report and close active issues.
Works on additional assignments as delegated by management.
Physical Demands
Weight Lifted: Up to 20 lbs, Occasionally 0-30% of time
Weight Carried: Up to 20 lbs, Occasionally 0-30% of time
Vision: Moderate, Frequently 31-65% of time
Kneeling/Stooping/Bending: Occasionally 0-30%
Standing/Walking: Frequently 31-65%
Pushing/Pulling: Occasionally 0-30%
Intensity of Work: Occasionally 0-30%
Job Requires: Reading, Writing, Reasoning, Talking, Keyboarding, Driving
Working at NGHS means being part of something special: a team invested in you as a person, an employee, and in helping you reach your goals.
NGHS: Opportunities start here.
Northeast Georgia Health System is an Equal Opportunity Employer and will not tolerate discrimination in employment on the basis of race, color, age, sex, sexual orientation, gender identity or expression, religion, disability, ethnicity, national origin, marital status, protected veteran status, genetic information, or any other legally protected classification or status.
$91k-119k yearly est. Auto-Apply 4d ago
Data Integration Engineer - Healthcare Data Infrastructure
Qualified Disability Specialists 3.9
Remote
Transform healthcare with us.
Qualified Health is seeking a Data Integration Engineer to serve as the technical implementation specialist for our healthcare data integration initiatives. In this hands-on role, you'll design and build robust data pipelines that transform raw healthcare data from diverse sources (Epic, LIMS, PACS, SharePoint, etc.) into production-ready datasets powering our AI platform. You'll work in close partnership with a Data Integration Manager who handles client relationships and program coordination, allowing you to focus on solving complex technical challenges, ensuring data quality, and building reusable integration patterns. Your roles are complementary:
You own: Technical implementation, ETL development, data quality validation, pipeline construction, troubleshooting, and production deployment execution
Manager owns: Partner relationships, requirements gathering, timeline management, stakeholder communication, issue escalation, and ensuring delivery meets expectations
Together you deliver: Successful data integrations that meet partner needs on time with high quality
This is a technical role for someone who loves working with data, enjoys solving puzzles, and takes pride in building reliable, production-grade solutions.
Key Responsibilities:
Technical Implementation & Development (80%)
Design and build ETL pipelines using PySpark, SQL, and Azure data services to process healthcare data from multiple source systems
Execute data extraction and transformation operations on complex healthcare datasets, ensuring accuracy and compliance with established standards
Develop data quality validation frameworks to identify and resolve issues during integration, QC, and backtesting phases
Troubleshoot technical issues including data schema mismatches, transformation logic errors, and performance bottlenecks
Build reusable data components and standardized integration patterns that accelerate future implementations
Optimize pipeline performance for large-scale healthcare datasets, ensuring efficient processing and resource utilization
Implement data validation rules specific to healthcare contexts (e.g., clinical code validation, temporal logic checks, referential integrity)
Write and maintain technical documentation for data pipelines, transformations, and integration patterns
Support production deployments by coordinating with infrastructure teams and conducting final testing
Collaboration & Problem-Solving (20%)
Partner with Data Integration Manager to translate partner requirements into technical specifications
Participate in technical discussions with partner IT teams to understand data schemas, access methods, and integration constraints
Provide technical guidance on data mapping specifications and transformation approaches
Identify data quality issues and work with Manager to coordinate resolution with partners
Share technical findings from QC and backtesting with Manager to inform partner conversations
Contribute to continuous improvement of tools, processes, and technical standards
Required Qualifications:
5+ years of experience in data analytics, dataengineering, or solution delivery roles, with demonstrated expertise in data integration and ETL processes
Strong analytical toolkit with proficiency in:
PySpark for distributed data processing
Advanced SQL for data querying and transformation
Excel for data analysis and reporting
Production ETL experience: Track record of building and maintaining production-grade data pipelines with proper error handling and monitoring
Data quality focus: Experience implementing validation frameworks and troubleshooting data quality issues
Healthcare data experience: Prior work with healthcare datasets (EHR, claims, clinical, lab data)
Problem-solving mindset: Ability to independently diagnose and resolve complex technical issues
Attention to detail: Commitment to accuracy, testing, and delivering reliable solutions
Collaborative working style: Comfortable partnering with non-technical colleagues and adapting to feedback
Bachelor's degree in Computer Science, Engineering, Data Science, Mathematics, or related technical field
Preferred Skills:
Epic Clarity experience: Direct work with Epic's relational database structure and clinical data models
Healthcare data standards knowledge: Understanding of FHIR, HL7v2, DICOM, LOINC, SNOMED, ICD-10
Azure cloud platform: Hands-on experience with Azure Databricks, Data Factory, Blob Storage, Delta Lake
Healthcare compliance awareness: Understanding of HIPAA requirements and healthcare data security best practices
Data warehouse/lakehouse experience: Familiarity with dimensional modeling and modern data architecture patterns
DevOps practices: Experience with Git, CI/CD pipelines, and infrastructure-as-code
Performance tuning: Proven ability to optimize complex data transformations for scale
LIMS/PACS experience: Prior work integrating laboratory or imaging systems data
Multiple data format fluency: Experience with JSON, XML, Parquet, CSV, and other healthcare interchange formats
Technical Environment:
Our data infrastructure is built on modern cloud technologies including:
Azure Databricks + Data Factory (plus Fabric and Snowflake integrations)
PySpark for distributed data processing
GitHub Actions + Terraform for CI/CD and Infrastructure as Code
Python with type-safe patterns and modern frameworks
Healthcare data formats including FHIR, Epic Clarity, and other EHR schemas
What Success Looks Like:
High-quality data pipelines delivered on schedule with thorough testing and documentation
Proactive issue identification with technical problems caught and resolved before impacting partners
Reusable components that reduce implementation time for subsequent integrations
Clean production deployments with minimal post-launch issues
Technical credibility with partner IT teams based on quality of work
Efficient troubleshooting with quick diagnosis and resolution of data quality issues
Impact & Growth Opportunity:
As a Data Integration Engineer at Qualified Health, you'll build the data infrastructure that powers AI-driven insights for major health systems. Your work directly enables better patient care by ensuring high-quality, reliable data flows into clinical decision support tools. This role offers deep technical learning in healthcare data, exposure to diverse health system architectures, and growth potential into senior technical or platform architecture roles as we scale.
Why Join Qualified Health?
This is an opportunity to join a fast-growing company and a world-class team, that is poised to change the healthcare industry. We are a passionate, mission-driven team that is building a category-defining product. We are backed by premier investors and are looking for founding team members who are excited to do the best work of their careers.
Our employees are integral to achieving our goals so we are proud to offer competitive salaries with equity packages, robust medical/dental/vision insurance, flexible working hours, hybrid work options and an inclusive environment that fosters creativity and innovation.
Our Commitment to Diversity
Qualified Health is an equal opportunity employer. We believe that a diverse and inclusive workplace is essential to our success, and we are committed to building a team that reflects the world we live in. We encourage applications from all qualified individuals, regardless of race, color, religion, gender, sexual orientation, gender identity or expression, age, national origin, marital status, disability, or veteran status.
Pay & Benefits: The pay range for this role is between $130,000 and $180,000, and will depend on your skills, qualifications, experience, and location. This role is also eligible for equity and benefits.
Join our mission to revolutionize healthcare with AI.
To apply, please send your resume through the application below.
$130k-180k yearly Auto-Apply 3d ago
Senior Data Engineer - New Solutions
Qventus 4.1
Remote
On this journey for over 12 years, Qventus is leading the transformation of healthcare. We enable hospitals to focus on what matters most: patient care. Our innovative solutions harness the power of machine learning, generative AI, and behavioral science to deliver exceptional outcomes and empower care teams to anticipate and resolve issues before they arise.
Our success in rapid scale across the globe is backed by some of the world's leading investors. At Qventus, you will have the opportunity to work with an exceptional, mission-driven team across the globe, and the ability to directly impact the lives of patients. We're inspired to work with healthcare leaders on our founding vision and unlock world-class medicine through world-class operations. #LI-JB1
Qventus is looking for a Senior DataEngineer to pioneer new data solutions at Qventus. Our Data team ensures that Qventus data users have the data they need to build and drive new and existing Qventus products at scale and cost. This includes bidirectional integration with hospital EMR sources via multiple channels (eg. FHIR), complex highly secure (HIPAA) transformations capable of normalizing information across various workflows and client nuances, integration with & discovery of multiple third party datasets from customer data to big-data claims, and more. Our products span the machine learning based orchestrations, real time hospital reporting, analytical insights, and interactive applications needed to improve the lives of patients and doctors across the country.
As a Senior DataEngineer, you will lead the design, development, and management across investments data assets & data pipelines. You will identify, monitor, and lead initiatives to collect, explore, and model data in close partnership with our data modeling lead and integrations partners. You will work closely with solution experts to design, iterate, and develop strategic enhancements to our Qventus foundational data model. You will be adept in partnering with cross-functional partners and data users to translate functional needs into data requirements and quality checks. You will be data curious and excited to have an impact on the team and in the company and to improve the quality of healthcare operations.
Key Responsibilities
Spearhead the discovery, evaluation, and integration of new datasets, collaborating (incl. pipeline development and data modeling/documentation) working closely with key data stakeholders to understand their impact and relevance to our core products and the healthcare domain.
Facilitate the technical management of data assets - clearly tracking and maintaining context on the data within the dataset lifecycle and sustaining tight partnerships with immediate partners on ingestion & solution dataengineering
Translate product / analytical vision into highly functional data pipelines supporting high quality & highly trusted data products (incl. designing data structures, building and scheduling data transformation pipelines, improving transparency etc.).
Set the standard for dataengineering practices within the company, guiding the architectural approaches, data pipeline designs, and the integration of cutting-edge technologies to foster a culture of innovation and continuous improvement.
What We're Looking For
Excellent cross-functional communication - the ability to break down complex technical components for technical and non-technical partners alike
Innate aptitude for interpreting complex datasets with demonstrated ability to discern underlying patterns, identify anomalies, and extract meaningful insights, demonstrating advanced data intuition and analytical skills. (Healthcare experience preferred)
Excellence in quality data pipeline design, development, and optimization to create reliable, modular, secure data foundations for the organization's data delivery system from applications to analytics & ML
Proven ability to independently handle ambiguous project requirements and lead data initiatives from start to finish, while collaborating extensively with cross-functional, non-technical teams to inform and shape product development.
Bonus Points For
5+ years of experience designing, building, and operating cloud-based, highly available, observable, and scalable data platforms utilizing large, diverse data sets in production to meet ambiguous business needs
Relevant industry certifications in a variety of Data Architecture services (Databricks Certified DataEngineer Professional, AWS Certified DataEngineer or Solutions Architect Professional, SnowPro Advanced Architect/DataEngineer, Microsoft Fabric DataEngineer, and dbt Analytics Engineering).
Experience with MLOps and/or developing and maintaining machine learning models and infrastructure
Experience with data visualization tools and analytics technologies (Sigma, Looker, Tableau, etc.)
Degree in Computer Science, Engineering, or related field
Compensation for this role is based on market data and takes into account a variety of factors, including location, skills, qualifications, and prior relevant experience. Salary is just one part of the total rewards package at Qventus. We also offer a range of benefits and perks, including Open Paid Time Off, paid parental leave, professional development, wellness and technology stipends, a generous employee referral bonus, and employee stock option awards.
Salary Range$150,000-$180,000 USD
Qventus values diversity in its workforce and proudly upholds the principles of Equal Opportunity Employment . We welcome all qualified applicants and ensure fair consideration for employment without discrimination based on any legally protected characteristics, including, but not limited to: veteran status, uniformed service member status, race, color, religion, sex, sexual orientation, gender identity, age, pregnancy (including childbirth, lactation and related medical conditions), national origin or ancestry, citizenship or immigration status, physical or mental disability, genetic information (including testing and characteristics) or any other category protected by federal, state or local law (collectively, "protected characteristics"). Our commitment to equal opportunity employment applies to all persons involved in our operations and prohibits unlawful discrimination by any employee, including supervisors and co-workers.
Qventus participates in the E-Verify program as required by law and is committed to providing reasonable accommodations to individuals with disabilities in compliance with Americans with Disabilities Act (ADA). In compliance with the California Consumer Privacy Act (CCPA), Qventus provides transparency into how applicant data is processed during the application process. Candidate information will be treated in accordance with our candidate privacy notice.
*Benefits and perks are subject to plan documents and may change at the company's discretion.
*Employment is contingent upon the satisfactory completion of our pre-employment background investigation and drug test.
$150k-180k yearly Auto-Apply 30d ago
Sr Data Engineer - Tampa
Photon Group 4.3
Remote
Responsibilities
Big Data - Spark, Hive, Java, CDP, 6 Years + with 3 Years of Dev experience on Big Data
Analyze data requirements and identify disparate data sources required for consolidation and distribution.
Document functional specifications and coordinate delivery of the same with technology team.
Review logical and conceptual data models in alignment with business requirements.
Work with the stakeholders to understand and gather requirements and produce business specifications
Validate solution implementations and ensure they meet business and functional requirements.
Provide production deployment support and investigate data quality issues.
Work with various technology leads to ensure the gaps in the data completeness or accuracy are bridged.
Qualifications
Subject matter expertise in financial industry - wholesale loans /lending business OR Capital Markets or Finance or Risk Reporting
Strong hands on experience with database and SQL is required.
Excellent documentation, analytical skills to produce process flow diagrams, business modelling, and functional design.
Proficiency in MS Office (Word, Excel, Visio, PowerPoint) with extensive experience using Excel for data analysis.
Experience with Data tracing/ Lineage efforts
Knowledge of logical and physical data model
Compensation, Benefits and Duration
Minimum Compensation: USD 42,000
Maximum Compensation: USD 148,000
Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role.
Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees.
This position is not available for independent contractors
No applications will be considered if received more than 120 days after the date of this post
$95k-138k yearly est. Auto-Apply 60d+ ago
Java backend data engineer - Smithfield RI
Photon Group 4.3
Remote
Looking for a highly motivated DataEngineer responsible for designing, building, and maintaining the operational and analytical capability in FI's data platforms. The work involves solution design, data analysis, end to end development, production rollout, and production support activities using Java Spring Batch, Python, AWS, Oracle, and Snowflake. The role will be involved in projects centered around maintaining and improving the quality of financial product information brought into our data ecosystem that will provide value to our compensation, sales, and reporting users. The role will be positioned as a more senior developer within a single squad.
The skillset aligns with a Java backend dataengineer. Top 3 required skills are Spring Batch, AWS, and SQL (Oracle, Snowflake).
The Expertise and Skills You Bring
Bachelor's or Master's Degree in a technology related field (e.g. Engineering, Computer Science, etc.) required with 5+ years of working experience
3+ years of Java development utilizing Spring frameworks. Experience writing batch jobs with Spring Batch is a must
2+ years of experience developing batch / backend applications that run in AWS, with focus on AWS Batch and S3
3+ years working with SQL (ANSI SQL, Oracle, Snowflake)
1+ years of Python development
Experience with Unix shell scripting (bash, ksh) and scheduling / orchestration tools (Control-M)
Strong data modeling skills with experience working with 3NF and Star Schema data models
Proven data analysis skills; not afraid to work in a complex data ecosystem
Hands-on experience on SQL query optimization and tuning to improve performance is desirable
Experience with DevOps, Continuous Integration and Continuous Delivery (Jenkins, Terraform, CloudFormation)
Experience in Agile methodologies (SCRUM)
Experience building and deploying containerized applications using Docker
Work experience in the financial services industry is a plus
Proven track record to handle ambiguity and work in a fast-paced environment, either independently or in a collaborative manner
Good interpersonal skills to work with multiple teams within the business unit and across the organization
The Value You Deliver
Passion and intellectually curiosity to learn new technologies and business areas
Working with our business partners to enhance their user experience and provide them curated data to help them drive business results more effectively
Enhancing data quality of financial product information in our systems that has large impact on sales, compensation, and reporting business users
Compensation, Benefits and Duration
Minimum Compensation: USD 40,000
Maximum Compensation: USD 142,000
Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role.
Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees.
This position is not available for independent contractors
No applications will be considered if received more than 120 days after the date of this post
$95k-138k yearly est. Auto-Apply 24d ago
SPARK Data Reconciliation Engineer- Jersey City
Photon Group 4.3
Remote
Job Title: PySpark Data Reconciliation Engineer
We're seeking a skilled PySpark Data Reconciliation Engineer to join our team and drive the development of robust data reconciliation solutions within our financial systems. You will be responsible for designing, implementing, and maintaining PySpark-based applications to perform complex data reconciliations, identify and resolve discrepancies, and automate data matching processes. The ideal candidate possesses strong PySpark development skills, experience with data reconciliation techniques, and the ability to integrate with diverse data sources and rules engines.
Key Responsibilities:
Data Reconciliation Development:
Design, develop, and test PySpark-based applications to automate data reconciliation processes across various financial data sources, including relational databases, NoSQL databases, batch files, and real-time data streams.
Implement efficient data transformation, matching algorithms (deterministic and heuristic) using PySpark and relevant big data frameworks.
Develop robust error handling and exception management mechanisms to ensure data integrity and system resilience within Spark jobs.
Data Analysis and Matching:
Collaborate with business analysts and data architects to understand data requirements and matching criteria.
Analyze and interpret data structures, formats, and relationships to implement effective data matching algorithms using PySpark.
Work with distributed datasets in Spark, ensuring optimal performance for large-scale data reconciliation.
Rules Engine Integration:
Integrate PySpark applications with rules engines (e.g., Drools) or equivalent to implement and execute complex data matching rules.
Develop PySpark code to interact with the rules engine, manage rule execution, and handle rule-based decision-making.
Problem Solving and Gap Analysis:
Collaborate with cross-functional teams to identify and analyze data gaps and inconsistencies between systems.
Design and develop PySpark-based solutions to address data integration challenges and ensure data quality.
Contribute to the development of data governance and quality frameworks within the organization.
Qualifications and Skills:
Bachelor's degree in Computer Science or a related field.
5+ years of hands-on experience in big data development, preferably with exposure to data-intensive applications.
Strong understanding of data reconciliation principles, techniques, and best practices.
Proficiency in PySpark, Apache Spark, and related big data technologies for data processing and integration.
Experience with rules engine integration and development
Strong analytical and problem-solving skills, with the ability to translate business requirements into technical solutions.
Excellent communication and collaboration skills to work effectively with business analysts, data architects, and other team members.
Familiarity with data streaming platforms (e.g., Kafka, Kinesis) and big data technologies (e.g., Hadoop, Hive, HBase) is a plus.
Compensation, Benefits and Duration
Minimum Compensation: USD 43,000
Maximum Compensation: USD 151,000
Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role.
Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees.
This position is not available for independent contractors
No applications will be considered if received more than 120 days after the date of this post
$95k-138k yearly est. Auto-Apply 60d+ ago
Sr Data Engineer - New Jersey
Photon Group 4.3
Remote
We need a Senior DataEngineer 10+ years exp proficient in Spark, Scala/Java, and Hive, with extensive hands-on development experience in the Big Data Ecosystem.
Key Responsibilities:
Design, implement, and optimize highly performant data pipelines using Spark, Scala/Java, and Hive on platforms like Cloudera Data Platform (CDP) or other Hadoop echo systems.
Take complete ownership of complex dataengineering projects within the big data ecosystem, covering the entire lifecycle from initial design and development to deployment and ongoing maintenance.
Develop robust and efficient Hive queries for extensive data analysis and reporting.
Champion and enforce best practices and coding standards for new and existing data flows to ensure they are robust, scalable, secure, and maintainable using Spark, Scala/Java, and Hive within the big data ecosystem.
Diagnose, troubleshoot, and resolve complex issues related to Spark, Scala/Java, and Hive applications and YARN resource management, implementing performance optimization solutions.
Proactively collaborate with stakeholders, working closely to develop solutions with full commitment and accountability.
Technical Skills & Experience:
Proven hands-on development expertise with Apache Spark
Strong programming proficiency in Scala and/or Java
In-depth knowledge and practical experience with Hive, including query optimization and data analysis.
Experience with data platforms such as Cloudera Data Platform (CDP) is highly desirable.
Education:
Bachelor's / Master's degree/University degree or equivalent experience
$95k-138k yearly est. Auto-Apply 60d+ ago
Senior Data Engineer - Dallas/California, United States
Photon Group 4.3
Remote
Greetings Everyone
Who are we?
For the past 20 years, we have powered many Digital Experiences for the Fortune 500. Since 1999, we have grown from a few people to more than 4000 team members across the globe that are engaged in various Digital Modernization. For a brief 1 minute video about us, you can check *****************************
Senior DataEngineer:
As part of the Mail Analytics DataEngineering team, you will be working on large-scale batch pipelines, data serving, data lakehouse, and analytics systems, enabling mission critical decision making, downstream, AI-powered capabilities, and more.
If you're passionate about building data infrastructure and platforms that power modern Data- and AI-driven business at scale, we want to hear from you!
Your Day
● Partner with Data Science, Product, and Engineering to collect requirements to define the data ontology for Mail Data & Analytics
● Lead and mentor junior DataEngineers to support Yahoo Mail's ever-evolving data needs
● Design, build, and maintain efficient and reliable batch data pipelines to populate core data sets
● Develop scalable frameworks and tooling to automate analytics workflows and streamline users interactions with data products
● Establish and promote standard methodologies for data operations and lifecycle management
● Develop new or improve and maintain existing large-scale data infrastructures and systems for data processing or serving, optimizing complex code through advanced algorithmic concepts and in-depth understanding of underlying data system stacks
● Create and contribute to frameworks that improve the efficacy of the management and deployment of data platforms and systems, while working with data infrastructure to triage and resolve issues
● Prototype new metrics or data systems
● Define and manage Service Level Agreements for all data sets in allocated areas of ownership
● Develop complex queries, very large volume data pipelines, and analytics applications to solve analytics and dataengineering problems
● Collaborate with engineers, data scientists, and product managers to understand business problems, technical requirements to deliver data solutions
● Engineering consulting on large and complex data lakehouse data
You Must Have
● BS in Computer Science/Engineering, relevant technical field, or equivalent practical experience, with specialization in DataEngineering
● 8+ years of experience building scalable ETL pipelines on industry standard ETL orchestration tools (Airflow, Composer, Oozie) with deep expertise in SQL, PySpark, or scala.
● 3+ years leading dataengineering development directly with business or data science partners
● Built, scaled, and maintained Multi-Terabyte data sets and having an expansive toolbox for debugging and unblocking large scale analytics challenges (skew mitigation, sampling strategies, accumulation patterns, data sketches, etc.)
● Experience with at least one major cloud's suite of offerings (AWS, GCP, Azure).
● Developed or enhanced ETL orchestrations tools or frameworks
● Worked within standard GitOps workflow (branch and merge, PRs, CI / CD systems)
● Experience working with GDPR
● Self-driven, challenge-loving, detail oriented, teamwork spirit, excellent communication skills, ability to multitask and manage expectations
Preferred
● MS/PhD in Computer Science/Engineering or relevant technical field, with specialization in DataEngineering
● 3 years experience in Google Cloud Platform technologies (BiqQuery, Dataproc, Dataflow, Composer, Looker)
Compensation, Benefits and Duration
Minimum Compensation: USD 58,000
Maximum Compensation: USD 203,000
Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role.
Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees.
This position is available for independent contractors
No applications will be considered if received more than 120 days after the date of this post
$95k-138k yearly est. Auto-Apply 22d ago
Data Engineer- Smithfield, RI
Photon Group 4.3
Remote
We are seeking a skilled and passionate DataEngineer to join our team and play a vital role in building and maintaining our data infrastructure. The ideal candidate will have extensive experience with AWS cloud services, particularly EMR, and strong proficiency in Spark and PySpark for data processing and transformation. This role will focus on designing, developing, and optimizing data pipelines to support our growing data needs.
Responsibilities:
AWS Data Services:
Design, implement, and manage data solutions on AWS, leveraging services such as EMR, S3, Glue, and others.
Optimize AWS data infrastructure for performance, scalability, and cost-effectiveness.
Implement best practices for data security and compliance on AWS.
Apache Spark & PySpark:
Develop and maintain scalable data pipelines using Apache Spark and PySpark.
Perform data extraction, transformation, and loading (ETL/ELT) processes.
Optimize Spark jobs for performance and efficiency.
Develop and maintain data quality checks and validation processes.
Amazon EMR:
Configure and manage EMR clusters for large-scale data processing.
Troubleshoot and resolve EMR cluster issues.
Optimize EMR cluster configurations for performance and cost.
Deploy and monitor spark applications on EMR.
Data Pipeline Development:
Design and implement robust and reliable data pipelines.
Automate data ingestion, processing, and storage processes.
Monitor data pipeline performance and troubleshoot issues.
Work with various data sources, both structured and unstructured.
Collaboration and Communication:
Collaborate with data scientists, analysts, and other engineers to understand data requirements.
Document data pipelines and infrastructure.
Communicate effectively with technical and non-technical stakeholders.
Participate in code reviews.
Performance Optimization:
Analyze query plans and optimize spark jobs.
Monitor and tune data processing performance.
Identify and resolve performance bottlenecks.
Qualifications:
Bachelor's degree in Computer Science, Data Science, or a related field (or equivalent experience).
Minimum 6-9 years of experience in a DataEngineering role.
Strong experience with Amazon Web Services (AWS) data services, particularly EMR.
Proficiency in Apache Spark and PySpark for data processing.
Experience with data warehousing and data lake concepts.
Strong SQL skills.
Experience with scripting languages (e.g., Python).
Understanding of data modeling and database design principles.
Experience with version control systems (e.g., Git).
Strong problem-solving and troubleshooting skills.
Excellent communication and collaboration skills.
Experience with other big data technologies (e.g., Hadoop, Hive, Kafka) is a plus.
Experience with data orchestration tools (ie airflow, step functions) is a plus.
Compensation, Benefits and Duration
Minimum Compensation: USD 37,000
Maximum Compensation: USD 130,000
Compensation is based on actual experience and qualifications of the candidate. The above is a reasonable and a good faith estimate for the role.
Medical, vision, and dental benefits, 401k retirement plan, variable pay/incentives, paid time off, and paid holidays are available for full time employees.
This position is not available for independent contractors
No applications will be considered if received more than 120 days after the date of this post
$95k-138k yearly est. Auto-Apply 60d+ ago
Data Engineer
FHI 4.4
Remote
It's fun to work in a company where people truly BELIEVE in what they're doing!
We're committed to bringing passion and customer focus to the business.
Ready to take the next step in your career journey?
We're committed to bringing passion and customer focus to the business. Ready to take the next step in your career journey?
FHI is hiring a junior to mid-level DataEngineer fluent in SQL and Python to build, debug, and maintain production data pipelines and system integrations. You'll own these pipelines end to end, partnering with the Business Intelligence team to deliver reliable data, actionable insights and process automation across the business. This role requires comfort reading, understanding, and improving existing systems, not just building greenfield solutions.
We use a combination of SQL Server, Snowflake, Fivetran, and Python with a heavy emphasis on the Pandas library. Experience with these exact technologies is nice but not required. If you have a solid SQL and Python foundation you will be able to pick up our specific tools quickly.
Key Responsibilities
Build and maintain ETL pipelines that ingest and validate source-system data with minimal transformation.
Design and implement SQL transformation layers that translate that raw source-system data into analyst-ready models.
Build and maintain data integrations via inbound and outbound APIs.
Independently troubleshoot data failures, across the entire data pipeline.
Automate manual processes and improve data delivery and reliability.
Create clear documentation (ETL processes, object usage, data models) and test/validate code changes.
Required Skills & Abilities
Strong SQL skills: able to encapsulate complex logic and messy data into simple, consistent models for analysts.
Practical experience with Python (Pandas nice to have).
Experience integrating external systems via inbound and outbound APIs.
Understanding of logging, error handling, and control flow required to operate production data pipelines.
Solid grasp of data architecture and modeling (normalized/denormalized, star/snowflake).
Experience using version control with a team, ideally Git.
Excellent analytical thinking, problem solving, and communication skills.
Proven ability to work independently, manage priorities, and deliver in a rapidly changing environment.
Education & Experience
Bachelor's in Computer Science, or equivalent experience. Non-traditional candidates are welcome; feel free to share a github link to a project you're proud of.
Experience maintaining data pipelines and integrations across SQL Server/Snowflake or similar environments.
Experience supporting data ingestion from enterprise systems (i.e. Workday, other ERP or HR platforms) and delivering data for downstream reporting.
Experience with BI or reporting platforms. We primarily use PowerBI, SSRS, and Workday Reports, but if you have strong fundamentals matter more than specific tool experience.
Relevant certifications are nice to have but not required.
Physical Requirements
Prolonged periods of sitting and working on a computer.
Location: Remote (U.S.) role with working hours aligned to Eastern Time (ET). Candidates in Central Time are welcome if able to work an ET schedule. Strong preference for candidates based in North Carolina.
Compensation:
Determined by experience, skills, and role fit.
By submitting this application, you are providing consent for FHI to contact you via phone (call or text) or email. This job description is not designed to cover or contain a comprehensive listing of activities, duties or responsibilities that are required of the associate for the job. Duties, responsibilities, and/or activities may change at any time with or without notice. FHI provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws. Additional Location: If you like wild growth and working with happy, enthusiastic over-achievers, you'll enjoy your career with us!
Additional Location:
If you like wild growth and working with happy, enthusiastic over-achievers, you'll enjoy your career with us!
$95k-139k yearly est. Auto-Apply 39d ago
Principal Data Engineer
Eye Care Partners 4.6
Ballwin, MO jobs
We are seeking an experienced professional who will serve as the Principal DataEngineer on our Data Platforms & Insights team. The Principal DataEngineer serves as a senior technical leader within the Data Platforms & Insights team, responsible for architecting, developing, and maintaining scalable data solutions that support enterprise-wide analytics, reporting, and data management initiatives. This role drives the design and implementation of robust data pipelines, ensures data quality and governance, and enables self-service analytics through a "Data as a Service" model. The Principal DataEngineer collaborates closely with cross-functional teams, business stakeholders, and third-party service providers to deliver high-impact data solutions, while also mentoring and supervising DataEngineers to uphold engineering standards and best practices.
ESSENTIAL DUTIES AND RESPONSIBILITIES
* Design, develop, and maintain scalable and efficient data pipelines using ETL tools and programming languages
* Develop integration solutions leveraging APIs to enable seamless communication between systems.
* Analyze data elements from various systems, data flow, dependencies, relationships and assist in designing conceptual physical and logical data models
* Implement data solutions across on-prem and cloud environments, ensuring performance, reliability, and scalability
* Ensure all data pipelines follow established data governance rules for data quality and completeness
* Maintain and evolve existing monitoring, logging, and alerting frameworks for proactively managing and troubleshooting data pipelines
* Manage source code repositories and deployment processes using modern tools
* Utilize Infrastructure as Code (IaC) tools to automate and manage infrastructure provisioning
* Work within Agile development framework to understand and transform business requirements into scalable and manageable solutions
* Work with various business and technical stakeholders and assist with data-related technical needs and issues
* Partner with leadership to define and evolve the long-term data architecture and engineering strategy, ensuring alignment with business goals
* Present solutions and options to leadership, project teams and other stakeholders adapting style to both technical and non-technical audiences
* Establish and enforce documentation standards for data pipelines, schemas, and infrastructure
* Ensures dataengineers and other technical teams adhere to documented design and development patterns and standards
* Conduct code reviews and provide guidance to other developers, fostering growth and development within the team
* Proactively monitor and resolve on-going production issues data pipelines, databases, and infrastructure
* Educate organization on latest trends and technologies in dataengineering, APIs, and streaming data
* Lead team on establishing industry best practices in dataengineering to ensure high-quality deliverables
* Adheres to all safety policies and procedures in performing job duties and responsibilities while supporting a culture of high quality and great customer service.
* Performs other duties that may be necessary or in the best interest of the organization.
QUALIFICATIONS
* Demonstrated ability to work efficiently and effectively in a fast-paced, matrixed environment, and ability to execute despite ambiguity
* Previous experience with a Healthcare company preferred
* Enjoys learning new technologies and systems
* Exhibits a positive attitude and is flexible in accepting work assignments and priorities
* Interpersonal skills to support customer service, functional, and teammate support need
* Knowledge of state and federal regulations for this position; general understanding of HIPAA guidelines
SUPERVISORY RESPONSIBILITIES
* Directly supervises DataEngineers on the Data Platforms & Insights team
* Carries out supervisory responsibilities in accordance with the organization's policies and applicable laws.
* Responsibilities include interviewing, hiring, and training employees, planning, assigning, and directing work; appraising performance, rewarding and disciplining employees, addressing complaints and resolving problems.
EDUCATION AND/OR EXPERIENCE
* Minimum Required: B.S. or B.A. Preferred in STEM (Science, Technology, Engineering, Math) field
* Minimum Required: 10+ years of hands-on-experience in the design, development, and implementation of data solutions
LICENSES AND CREDENTIALS
* Minimum Required: None
SYSTEMS AND TECHNOLOGY
* Proficient in Microsoft Excel, Word, PowerPoint, Outlook
* Experience working with the following:
* Snowflake development and support
* Advanced SQL knowledge with strong query writing skills
* Object-oriented/object function scripting languages: Python, Java, Scala, etc.
* AWS cloud services: EC2, EMR, RDS, DMS
* Relational databases such as SQL Server and object relational databases such as PostgreSQL
* Data analysis, ETL, and workflow automation
* Multiple ETL/ELT tools and cloud-based data hubs such as Fivetran
* Stream-processing systems: Kafka, Spark-Streaming, etc
* Source code management and deployment tools (e.g., Git, Jenkins, dbt, Docker).
* Infrastructure as Code (IaC) tools (e.g., Terraform, Ansible, CloudFormation)
* Enterprise MDM solutions
LOCATION
* This position is located in St Louis, Missouri and offers a hybrid work schedule. Candidates living in Alabama, Arizona, Florida, Georgia, Illinois, Indiana, Kansas, Kentucky, Michigan, Minnesota, Missouri, New Jersey, N. Carolina, Ohio, Oklahoma, Pennsylvania, Texas and Virginia may also be considered for remote work.
If you need assistance with this application, please contact **************. Please do not contact the office directly - only resumes submitted through this website will be considered.
EyeCare Partners is an equal opportunity/affirmative action employer. All applicants will be considered for employment without attention to race, color, religion, sex, sexual orientation, gender identity, national origin, veteran or disability status.
$77k-96k yearly est. Auto-Apply 60d+ ago
Principal Data Engineer - ML Platforms
Altarum 4.5
Remote
Altarum | Data & AI Center of Excellence (CoE) Altarum is building the future of data and AI infrastructure for public health - and we're looking for a Principal DataEngineer - ML Platforms to help lead the way. In this cornerstone role, you will design, build, and operationalize the modern data and ML platform capabilities that power analytics, evaluation, AI modeling, and interoperability across all Altarum divisions.
If you want to architect impactful systems, enable data science at scale, and help ensure public health and Medicaid programs operate with secure, explainable, and trustworthy AI - this role is for you. What You'll Work On
This role blends deep engineering with applied ML enablement: ML Platform Engineering: modern lakehouse architecture, pipelines, MLOps lifecycle Applied ML enablement: risk scoring, forecasting, Medicaid analytics NLP/Generative AI support: RAG, vectorization, health communications Causal ML operationalization: evaluation modeling workflows Responsible/Trusted AI engineering: model cards, fairness, compliance Your work ensures that Altarum's public health and Medicaid programs run on secure, scalable, reusable, and explainable data and AI infrastructure. What You'll Do
Platform Architecture & Delivery
Design and operate modern, cloud-agnostic lakehouse architecture using object storage, SQL/ELT engines, and dbt.
Build CI/CD pipelines for data, dbt, and model delivery (GitHub Actions, GitLab, Azure DevOps).
Implement MLOps systems: MLflow (or equivalent), feature stores, model registry, drift detection, automated testing.
Engineer solutions in AWS and AWS GovCloud today, with portability to Azure Gov or GCP.
Use Infrastructure-as-Code (Terraform, CloudFormation, Bicep) to automate secure deployments.
Pipelines & Interoperability
Build scalable ingestion and normalization pipelines for healthcare and public health datasets, including:
FHIR R4 / US Core (strongly preferred)
HL7 v2 (strongly preferred)
Medicaid/Medicare claims & encounters (strongly preferred)
SDOH & geospatial data (preferred)
Survey, mixed-methods, and qualitative data
Create reusable connectors, dbt packages, and data contracts for cross-division use.
Publish clean, conformed, metrics-ready tables for Analytics Engineering and BI teams.
Support Population Health in turning evaluation and statistical models into pipelines.
Data Quality, Reliability & Cost Management
Define SLOs and alerting; instrument lineage & metadata; ensure ≥95% of data tests pass.
Perform performance and cost tuning (partitioning, storage tiers, autoscaling) with guardrails and dashboards.
Applied ML Enablement
Build production-grade pipelines for risk prediction, forecasting, cost/utilization models, and burden estimation.
Develop ML-ready feature engineering workflows and support time-series/outbreak detection models.
Integrate ML assets into standardized deployment workflows.
Generative AI Enablement
Build ingestion and vectorization pipelines for surveys, interviews, and unstructured text.
Support RAG systems for synthesis, evaluation, and public health guidance.
Enable Palladian Partners with secure, controlled-generation environments.
Causal ML & Evaluation Engineering
Translate R/Stata/SAS evaluation code into reusable pipelines.
Build templates for causal inference workflows (DID, AIPW, CEM, synthetic controls).
Support operationalization of ARA's applied research methods at scale.
Responsible AI, Security & Compliance
Implement Model Card Protocol (MCP) and fairness/explainability tooling (SHAP, LIME).
Ensure compliance with HIPAA, 42 CFR Part 2, IRB/DUA constraints, and NIST AI RMF standards.
Enforce privacy-by-design: tokenization, encryption, least-privilege IAM, and VPC isolation.
Reuse, Shared-Services, and Enablement
Develop runbooks, architecture diagrams, repo templates, and accelerator code.
Pair with data scientists, analysts, and SMEs to build organizational capability.
Provide technical guidance for proposals and client engagements.
Your First 90 Days - You will make a meaningful impact fast. Expected outcomes include:
Platform skeleton operational: repo templates, CI/CD, dbt project, MLflow registry, tests.
Two pipelines in production (e.g., FHIR → analytics and claims normalization).
One end-to-end CoE lighthouse MVP delivered (ingestion → model → metrics → BI).
Completed playbooks for GovCloud deployment, identity/secrets, rollback, and cost control.
Success Metrics (KPIs)
Pipeline reliability meeting SLA/SLO targets.
≥95% data tests passing across pipelines.
MVP dataset onboarding ≤ 4 weeks.
Reuse of platform assets across ≥3 divisions.
Cost optimization and budget adherence.
What You'll Bring
7-10+ years in dataengineering, ML platform engineering, or cloud data architecture.
Expert in Python, SQL, dbt, and orchestration tools (Airflow, Glue, Step Functions).
Deep experience with AWS + AWS GovCloud.
CI/CD and IaC experience (Terraform, CloudFormation).
Familiarity with MLOps tools (MLflow, Sagemaker, Azure ML, Vertex AI).
Ability to operate in regulated environments (HIPAA, 42 CFR Part 2, IRB).
Preferred:
Experience with FHIR, HL7, Medicaid/Medicare claims, and/or SDOH datasets.
Databricks, Snowflake, Redshift, Synapse.
Event streaming (Kafka, Kinesis, Event Hubs).
Feature store experience.
Observability tooling (Grafana, Prometheus, OpenTelemetry).
Experience optimizing BI datasets for Power BI.
Logistical Requirements
At this time, we will only accept candidates who are presently eligible to work in the United States and will not require sponsorship.
Our organization requires that all work, for the duration of your employment, must be completed in the continental U.S. unless required by contract.
If you're near one of our offices (Arlington, VA; Silver Spring, MD; or Novi, MI), you'll join us in person one day every other month (6 times per year) for a fun, purpose-driven Collaboration Day. These days are filled with creative energy, meaningful connection, and team brainstorming!
Must be able to work during Eastern Time unless approved by your manager.
Employees working remotely must have a dedicated, ergonomically appropriate workspace free from distractions with a mobile device that allows for productive and efficient conduct of business.
Altarum is a nonprofit organization focused on improving the health of individuals with fewer financial resources and populations disenfranchised by the health care system. We work primarily on behalf of federal and state governments to design and implement solutions that achieve measurable results. We combine our expertise in public health and health care delivery with technology development and implementation, practice transformation, training and technical assistance, quality improvement, data analytics, and applied research and evaluation. Our innovative solutions and proven processes lead to better value and health for all.
Altarum is an equal opportunity employer that provides employment and opportunities to all qualified employees and applicants without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, or any other characteristic protected by applicable law.
$72k-98k yearly est. Auto-Apply 60d ago
Data Engineer Senior
Massachusetts Eye and Ear Infirmary 4.4
Somerville, MA jobs
Site: Mass General Brigham Incorporated
Mass General Brigham relies on a wide range of professionals, including doctors, nurses, business people, tech experts, researchers, and systems analysts to advance our mission. As a not-for-profit, we support patient care, research, teaching, and community service, striving to provide exceptional care. We believe that high-performing teams drive groundbreaking medical discoveries and invite all applicants to join us and experience what it means to be part of Mass General Brigham.
Job Summary
Summary
Responsible for designing, developing, and maintaining the data architecture and infrastructure within an organization. This position plays a crucial role in managing large-scale data systems and ensuring the efficient flow, storage, and accessibility of data for various stakeholders, such as data analysts, data scientists, and business users.
Essential Functions
-Design, develop, and implement data pipelines and ETL/ELT code to support business requirements.
-Work on cross-functional teams delivering enterprise solutions for internal and external clients.
-Assume ownership for delivering code revisions and enhancements from design through development and production installation.
-Maintain and optimize various components of the data pipeline architecture.
-Become subject matter expert for internal and external data products Ensure design solutions can scale and meet technical standards and performance benchmarks.
-Identify inefficient processes and develop recommendations and design solutions.
-Lead code review sessions to validate technical solutions and facilitate knowledge sharing.
Qualifications
Education
Bachelor's Degree Related Field of Study required
Can this role accept experience in lieu of a degree?
Yes
Principal Responsibilities
Works with cross-functional teams to understand functional product requirements and deliver on strategic data and analytics initiatives.
Design, build, test and maintain architectures within the ‘Quality Data Hub' of our cloud-based enterprise data warehouse.
Build ETL/ELT ingestions for OCMO related data and handle related monitoring and support duties.
Accrues advanced knowledge of the OCMO product domains and can quickly apply development strategies to meet the stakeholder requirements for the data products.
Define and implement the processes and tooling through which enterprise curated data models are built and maintained.
Partners with solutions architects, analytics engineers, and data visualization developers to understand data extraction and transformation needs and builds a platform and related processes suited to them.
Develop and enforce change management and versioning processes for code promotion.
In collaboration with lead analytic staff, architect and enforce thorough quality assurance, testing and peer review procedures to ensure the accuracy, reliability, and validity of end products.
Helps identify potential bottlenecks within the development lifecycle and propose high-level strategies to overcome them.
Triage, troubleshoot and resolve data issues from end users and internal team members.
Build and foster relationships with senior leadership/physicians and key program stakeholders to understand multifaceted business problems and develop analytical solutions to complex issues to satisfy reporting and analytical needs.
Supports other team members with promotion to production, ensures a consistent process across the team; informs other dataengineers and domain team leads of new changes to promotion process.
Guide and train dataengineers by familiarizing them with the data products they will use and providing them with relevant information about OCMO domains.
Follows MGB Digital guidelines and standard practices to leverage existing tools and methods to efficiently develop modeling solutions, ensure that work is not duplicated, and is appropriately transitioned between teams.
Uses the Mass General Brigham values to govern decisions, actions, and behaviors. These values guide how we get our work done: Patients, Affordability, Accountability & Service Commitment, Decisiveness, Innovation & Thoughtful Risk; and how we treat each other: Diversity & Inclusion, Integrity & Respect, Learning, Continuous Improvement & Personal Growth, Teamwork & Collaboration.
Other duties and responsibilities as assigned.
Knowledge, Skills and Abilities
Prior experience working with healthcare data is strongly preferred.
Prior experience working with dbt is strongly preferred.
Ability to function effectively and independently in a fast-paced environment, organize and prioritize work independently, and meet tight deadlines.
Ability to manage multiple projects simultaneously, set priorities, and collaborate with team members and others throughout the organization.
Possess strong interpersonal skills to effectively communicate with cross functional teams including staff at all levels of the organization.
Willing to contribute to and foster a team player culture where all are encouraged and willing to share information accurately.
Able and motivated to mentor/train junior staff members.
Knowledge of agile principles and experience working within an agile team is preferred.
Practical problem-solving abilities, i.e. the ability to formulate hypotheses, test options and move forward in a fast-paced environment.
Excellent interpersonal skills, including strong customer service orientation and the ability to translate complex technical concepts to non-technical audiences.
Advanced SQL DML skills required, with comparable experience in writing data functions (e.g. TSQL procs, Snowflake UDFs, etc.).
Experience
Bachelor's or master's degree in computer science, informatics, statistics, or related field and an interest in healthcare and the use of technology to support clinical care.
5+ years' experience with data modelling, ETL/ELT development, or similar role working with complex SQL queries and data extraction/transformation.
Experience with cloud data warehousing environments such as Microsoft Azure and Snowflake are a plus
Experience with ETL/data modelling tools such as dbt, Informatica, and Ab Initio are a plus.
Must possess a strong background in data warehousing projects.
Must be able to identify, triage, and resolve or dispatch issues.
Must possess strong data analysis skills and be able to perform data analysis using SQL, SAS, or similar query languages.
Must possess strong oral and written communication skills.
Must be capable of working independently with limited to no supervision.
Must be willing to contribute to and foster a team player culture where all are encouraged and willing to share information accurately.
Additional Job Details (if applicable)
Working Model Requirements
Hybrid with onsite work required in office, candidate must be flexible for weekly or monthly business needs
M-F Eastern business hours required
On remote workdays, employees must use a stable, secure, and compliant workstation in a quiet environment. Teams video is required and must be accessed using MGB-provided equipment.
Remote Type
Hybrid
Work Location
399 Revolution Drive
Scheduled Weekly Hours
40
Employee Type
Regular
Work Shift
Day (United States of America)
Pay Range
$93,953.60 - $136,739.20/Annual
Grade
7
At Mass General Brigham, we believe in recognizing and rewarding the unique value each team member brings to our organization. Our approach to determining base pay is comprehensive, and any offer extended will take into account your skills, relevant experience if applicable, education, certifications and other essential factors. The base pay information provided offers an estimate based on the minimum job qualifications; however, it does not encompass all elements contributing to your total compensation package. In addition to competitive base pay, we offer comprehensive benefits, career advancement opportunities, differentials, premiums and bonuses as applicable and recognition programs designed to celebrate your contributions and support your professional growth. We invite you to apply, and our Talent Acquisition team will provide an overview of your potential compensation and benefits package.
EEO Statement:
0100 Mass General Brigham Incorporated is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religious creed, national origin, sex, age, gender identity, disability, sexual orientation, military service, genetic information, and/or other status protected under law. We will ensure that all individuals with a disability are provided a reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. To ensure reasonable accommodation for individuals protected by Section 503 of the Rehabilitation Act of 1973, the Vietnam Veteran's Readjustment Act of 1974, and Title I of the Americans with Disabilities Act of 1990, applicants who require accommodation in the job application process may contact Human Resources at **************.
Mass General Brigham Competency Framework
At Mass General Brigham, our competency framework defines what effective leadership “looks like” by specifying which behaviors are most critical for successful performance at each job level. The framework is comprised of ten competencies (half People-Focused, half Performance-Focused) and are defined by observable and measurable skills and behaviors that contribute to workplace effectiveness and career success. These competencies are used to evaluate performance, make hiring decisions, identify development needs, mobilize employees across our system, and establish a strong talent pipeline.
$94k-136.7k yearly Auto-Apply 3d ago
ETL Architect - Snowflake and SSIS Experience
Northeast Georgia Medical Center 4.8
Georgia jobs
Job Category:
Administrative & Clerical
Work Shift/Schedule:
8 Hr Morning - Afternoon
Northeast Georgia Health System is rooted in a foundation of improving the health of our communities.
ETL Architect is a high-level business intelligence professional that is a key part of the Data and Analytics Team and will be generally responsible for the dimensional model design, end to end ETL architecture design, ETL development, and implementation of data models and development of the Data Migration and Data Warehousing solutions. The ETL Architect is also responsible for requirements analysis, source data profiling, identifying data quality issues, designing transformation logic to address data quality issues, experienced with the concepts and/or implementation of metadata data capture, data lineage, master data management and reference data management. The ETL architect designs for data warehouse/data mart solutions should include appropriate staging databases to handle change data capture on the source data, error handling and audit balance control mechanisms. The ETL Architect will have extensive interaction with operational team members and leadership, and will be tasked with analysis of business data and analytics requirements, and the transformation of those requirements into analytics solutions. Working with customers and Management to determine priorities as well mentoring of other team members on ETL concepts and strategies is also an integral role of the position. Working with Infrastructure and DBA team members, ETL Architect is also responsible for keeping the ETL Environments updated periodically to the latest version of the ETL software.
Minimum Job Qualifications
Licensure or other certifications: Epic Revenue or Clinical Data Model or Cogito Project Manager Certificate required (must be obtained within 6 months of date of hire).
Educational Requirements: Bachelors Degree in Computer Science or related field. In lieu of Bachelor's Degree related experience in the field, 11 years progressively responsible, professional job specific work experience and an equivalent combination of education and training that provides the required knowledge, skills and abilities, combined with demonstrated professional growth and achievement.
Minimum Experience: A minimum of seven (7) years experience as a various BI and other analytic reporting software applications with strong preference for EPIC, SQL, and Crystal Report Writing. Prior experience providing production level support and/or development in business intelligence (BI)/enterprise data warehouse (EDW) environments.
Other:
Preferred Job Qualifications
Preferred Licensure or other certifications:
Preferred Educational Requirements:
Preferred Experience: Experience in database development and management tasks such as creating indexes, functions, stored procedures, creating databases, file groups, users, configuring permissions and assigning roles to users. Working knowledge of data warehouse concepts, enterprise data management, modeling techniques and ability to create solutions, participate in strategy discussions and review them with customer as well as project stakeholders.
Other:
Job Specific and Unique Knowledge, Skills and Abilities
Strong analytical and interpersonal skills; demonstrated ability to work with management and customers; proven ability to work independently or as a team member; strong problem solving and troubleshooting skills; initiative in enhancing system productivity, functionality, and performance.
Advanced understanding of Data Migration/Data Warehousing/Data Mart delivery life cycle, from requirements definition to creating source to target mapping documents with transformation rules, Error handling design, data quality, design audit Balance control mechanisms to be able to automate jobs, notify proactively failures, write test cases and create templates to document all aspects of the projects throughout the lifecycle.
Strong experience with concepts of data profiling, identifying and measuring data quality issues, advanced experience with Master data and reference data management techniques to eliminate hard coding within the ETL processes/jobs.
Excellent communication (written, verbal and presentation) and people-facing skills, Proven ability to maintain positive peer, customer relationships, interacting positively and productively with teams across organizational lines. Demonstrates skills in leadership, critical thinking, and consensus and relationship building.
Ability to review, re-engineer and/or tune legacy SQL queries and/or ETL jobs/processes to make them run efficiently and within a load window.
Must be able to function in dynamic environment subject to changes in schedules and priorities, working on multiple projects/ issues concurrently. Responsibilities may require evening and weekend work in response to needs of the systems supported.
Experience in estimation, specification gathering, review code/designs, general planning for execution, monitoring and control.
Hands-on with ability to design, re-design (schemas, data models, entity relationships etc.), and probe and guide developers.
Possess a solid understanding of data warehousing concepts, including relational and multidimensional modeling.
Experience in designing, building, testing, executing and performance tuning of ETL and reporting applications, including automation of ETL jobs for recurring imports; fluent in Microsoft SQL.
Experience in designing and developing database objects, tables, stored procedures, views, and triggers.
Essential Tasks and Responsibilities
Designs end to end ETL architecture for Data Migration, Data Warehousing or Data Mart Solutions using Microsoft SSIS tool. Ensure all ETL Jobs have adequate error handling and audit balance control mechanisms implemented to enable automation.
Stands up a DW/ETL Center of Excellence to document and enforce best practices, standards, naming conventions etc. Creates a checklist to review all solutions developed to ensure adherence to the CoE standards. Mentors and trains staff to hold periodic CoE meetings.
Implement a reference data management system using tools like Microsoft Master Data Management Services to eliminate hardcoding in ETL jobs.
Maintains a service oriented approach with customers, providing follow up and responding to issues/ concerns in a timely manner. Answer staff and client questions in person, via email and via phone. Provide after-hours on-call support on a scheduled basis. Responsible for ensuring service level agreement adherence.
Works with staff to adhere to department/organization policies and procedures, compliance and regulatory agency requirements.
Works with leaders in IT as well as operations to identify data from other source systems to form richer data sets, including the capture of derived data to address gaps in source system repositories.
Promotes and utilizes the data request and development process, including the development of an understanding of a data request, assessment of the appropriate tool to be used to provide the data, creation of a requirements document in partnership with the operational stakeholder and application analyst, development of the data extract and validation of the extract by the operational stakeholder to ensure quality, timeliness and accountability.
Participates in and contributes to the data governance program - specifically, validating data lineage, capturing metadata for all existing and any newly developed assets (ETL Jobs), profile various data sources and address data quality issues, work with source system owners to reduce upstream data entry errors.
Assists with all upgrade, update or enhancement efforts associated with the ETL tool.
In partnership with the Manager of Business Intelligence and Analytics, serves as a mentor and advisor to the BI technical resources and other analysts and provides general project management skills to produce BI views that address customer's needs.
Builds and maintains SQL scripts, indexes and complex queries for data analysis and extraction.
Troubleshoots problems regarding the databases, applications, and development tools.
Assists with the development and implementation of new data sources into the data warehouse or determine ways to be able to utilize data in other systems to support data and analytics for the organization.
Conducts or coordinate tests to ensure that intelligence is consistent with defined needs.
Responsible for data collection/extraction and developing methods for database design and validation reports.
Prepares and maintains key documents and technical specifications for new and existing solutions.
Documents specifications for business intelligence or information technology (IT) reports, dashboards, or other outputs.
In collaboration with other team members and leadership, supports the evaluation, adoption and implementation of new methods, tools and technologies in order to meet current gaps in the organization-wide data and analytics infrastructure.
Works closely with end-users to identify and explain data and reporting discrepancies.
Prepares and maintains key documents and technical specifications for new and existing solutions.
Provides solution design, perform reviews and participate strategic discussions.
Provides application analysis and data modeling design to collect data or to extract data from various disparate systems.
Monitors incident management queue. Efficiently troubleshoot, resolve, report and close active issues.
Works on additional assignments as delegated by management.
Physical Demands
Weight Lifted: Up to 20 lbs, Occasionally 0-30% of time
Weight Carried: Up to 20 lbs, Occasionally 0-30% of time
Vision: Moderate, Frequently 31-65% of time
Kneeling/Stooping/Bending: Occasionally 0-30%
Standing/Walking: Frequently 31-65%
Pushing/Pulling: Occasionally 0-30%
Intensity of Work: Occasionally 0-30%
Job Requires: Reading, Writing, Reasoning, Talking, Keyboarding, Driving
Working at NGHS means being part of something special: a team invested in you as a person, an employee, and in helping you reach your goals.
NGHS: Opportunities start here.
Northeast Georgia Health System is an Equal Opportunity Employer and will not tolerate discrimination in employment on the basis of race, color, age, sex, sexual orientation, gender identity or expression, religion, disability, ethnicity, national origin, marital status, protected veteran status, genetic information, or any other legally protected classification or status.
$83k-107k yearly est. Auto-Apply 38d ago
Data & BI Engineer
L R S 4.3
Springfield, MO jobs
LRS is seeking a skilled Data & BI Engineer to design, build, and maintain scalable data solutions that power analytics and reporting across the organization. This hybrid role combines data architecture and engineering (integration, pipelines, modeling) with BI development (dashboards, visualizations, and insights). The ideal candidate is comfortable working across the full data stack and collaborating with business stakeholders to deliver actionable intelligence.
This is an in-office position based out of our headquarters in Springfield, Illinois.
Requirements
5+ years of experience in dataengineering and data analytics.
Proficiency in SQL Server and T-SQL.
Experience with data modeling (star/snowflake schemas) and ETL/ELT processes.
Experience with BI tools.
Strong understanding of data governance, security, and performance optimization.
Excellent communication and stakeholder engagement skills.
The following will make you a stronger candidate
Experience working with data warehouses.
Experience with Power BI, DAX, and Power Query.
Familiarity with Microsoft Fabric.
Key Responsibilities
Design and implement data pipelines using modern ETL/ELT tools.
Develop and maintain semantic data models optimized for reporting and analytics.
Build compelling Power BI dashboards and reports with functional, user-friendly visuals.
Collaborate with business units to understand data needs and translate them into technical solutions.
Ensure data quality, integrity, and governance across systems.
Optimize performance of data solutions and BI assets.
Support data integration across cloud and on-premises systems.
Document architecture, data flows, and reporting logic.
Success Factors
The successful candidate will demonstrate expertise across the data stack, delivering reliable, high-quality data solutions and actionable insights. Success in this role will be measured by your ability to collaborate with business stakeholders, optimize data-driven processes, and drive impactful analytics initiatives.
Organization Structure
The LRS IT team consists of a Chief Information Officer, Director of IT, Director of Applications, Director of Information Security, and teams for networking, infrastructure, cloud, communications, end-user services, and applications. The team is based in Springfield, IL and manages the global operations at LRS. You will report to the Chief Information Officer.
LRS is an equal opportunity employer. Applicants for employment will receive consideration without unlawful discrimination based on race, color, religion, creed, national origin, sex, age, disability, marital status, gender identity, domestic partner status, sexual orientation, genetic information, citizenship status or protected veteran status.
Salary Range: $90,000-$130,000. This salary range represents the low and high end for this position. The salary will vary depending on factors including experience and skills. The range listed is just one component of LRS' total employee compensation, as we have a generous benefits package.
$90k-130k yearly 60d+ ago
Data Platform Engineer
Monogram Health Inc. 3.7
Brentwood, TN jobs
Job DescriptionPosition:
Data Platform Engineer
The DataEngineering team is seeking a highly skilled and experienced Data Platform Engineer with expertise in DataEngineering, Database Modeling, and modern Cloud Data Platforms. The Data Platform Engineer designs, builds, and maintains scalable and secure data infrastructure, tools, and pipelines to support data analytics, machine learning, and business intelligence initiatives. They will also work closely with other teams, such as data analysts, data scientists, and software developers, to provide them with the data they need to perform their job functions.
Responsibilities
Design and implement robust, scalable, and efficient data models and pipelines across cloud-based platforms.
Develop, optimize, and maintain ETL/ELT pipelines using Azure Data Factory and Databricks.
Build and orchestrate Databricks Notebooks and Jobs using PySpark, Spark SQL, or Scala Spark.
Develop and manage data models, data warehousing solutions, and data integration architectures in Azure.
Implement Azure Functions, Azure WebApps, and Application Insights to support microservices and monitor distributed systems.
Configure and manage Databricks clusters, including autoscaling, Photon acceleration, and job orchestration.
Collaborate with cross-functional teams to support data-driven decision-making and analytics use cases.
Ensure data quality, governance, and security across the data lifecycle.
Collaborate with product managers by estimating technical tasks and deliverables.
Uphold the mission and values of Monogram Health in all aspects of your role and activities.
Position Requirements
A bachelor's degree in computer science, data science, software engineering or related field.
Minimum of five (5) years in designing and hands-on development in cloud-based analytics solutions, which includes a minimum of three (3) years' hands on work with big data frameworks and tools, such as Apache Kafka and Spark.
Expert level knowledge of Python or other scripting languages required.
Proficiency in SQL and other data query languages.
Understanding of data modeling and schema design principles
Ability to work with large datasets and perform data analysis
Designing and building data integration pipelines using API's and Streaming ingestion methods is desirable.
Familiarity with DevOps practices, including automation, CI/CD, and infrastructure as code (IaC).
Thorough understanding of Azure Cloud Infrastructure offerings.
Demonstrated problem-solving and troubleshooting skills.
Team player with demonstrated written and communication skills.
Benefits
Comprehensive Benefits - Medical, dental, and vision insurance, employee assistance program, employer-paid and voluntary life insurance, disability insurance, plus health and flexible spending accounts
Financial & Retirement Support - Competitive compensation, 401k with employer match, and financial wellness resources
Time Off & Leave - Paid holidays, flexible vacation time/PSSL, and paid parental leave
Wellness & Growth - Work life assistance resources, physical wellness perks, mental health support, employee referral program, and BenefitHub for employee discounts
Monogram Health is a leading multispecialty provider of in-home, evidence-based care for the most complex of patients who have multiple chronic conditions. Monogram health takes a comprehensive and personalized approach to a person's health, treating not only a disease, but all of the chronic conditions that are present - such as diabetes, hypertension, chronic kidney disease, heart failure, depression, COPD, and other metabolic disorders.
Monogram Health employs a robust clinical team, leveraging specialists across multiple disciplines including nephrology, cardiology, endocrinology, pulmonology, behavioral health, and palliative care to diagnose and treat health issues; review and prescribe medication; provide guidance, education, and counselling on a patient's healthcare options; as well as assist with daily needs such as access to food, eating healthy, transportation, financial assistance, and more. Monogram Health is available 24 hours a day, 7 days a week, and on holidays, to support and treat patients in their home.
Monogram Health's personalized and innovative treatment model is proven to dramatically improve patient outcomes and quality of life while reducing medical costs across the health care continuum.
$75k-103k yearly est. 10d ago
Data Platform Engineer
Monogram Health 3.7
Brentwood, TN jobs
Data Platform Engineer The DataEngineering team is seeking a highly skilled and experienced Data Platform Engineer with expertise in DataEngineering, Database Modeling, and modern Cloud Data Platforms. The Data Platform Engineer designs, builds, and maintains scalable and secure data infrastructure, tools, and pipelines to support data analytics, machine learning, and business intelligence initiatives. They will also work closely with other teams, such as data analysts, data scientists, and software developers, to provide them with the data they need to perform their job functions.
Responsibilities
* Design and implement robust, scalable, and efficient data models and pipelines across cloud-based platforms.
* Develop, optimize, and maintain ETL/ELT pipelines using Azure Data Factory and Databricks.
* Build and orchestrate Databricks Notebooks and Jobs using PySpark, Spark SQL, or Scala Spark.
* Develop and manage data models, data warehousing solutions, and data integration architectures in Azure.
* Implement Azure Functions, Azure WebApps, and Application Insights to support microservices and monitor distributed systems.
* Configure and manage Databricks clusters, including autoscaling, Photon acceleration, and job orchestration.
* Collaborate with cross-functional teams to support data-driven decision-making and analytics use cases.
* Ensure data quality, governance, and security across the data lifecycle.
* Collaborate with product managers by estimating technical tasks and deliverables.
* Uphold the mission and values of Monogram Health in all aspects of your role and activities.
Position Requirements
* A bachelor's degree in computer science, data science, software engineering or related field.
* Minimum of five (5) years in designing and hands-on development in cloud-based analytics solutions, which includes a minimum of three (3) years' hands on work with big data frameworks and tools, such as Apache Kafka and Spark.
* Expert level knowledge of Python or other scripting languages required.
* Proficiency in SQL and other data query languages.
* Understanding of data modeling and schema design principles
* Ability to work with large datasets and perform data analysis
* Designing and building data integration pipelines using API's and Streaming ingestion methods is desirable.
* Familiarity with DevOps practices, including automation, CI/CD, and infrastructure as code (IaC).
* Thorough understanding of Azure Cloud Infrastructure offerings.
* Demonstrated problem-solving and troubleshooting skills.
* Team player with demonstrated written and communication skills.
Benefits
* Comprehensive Benefits - Medical, dental, and vision insurance, employee assistance program, employer-paid and voluntary life insurance, disability insurance, plus health and flexible spending accounts
* Financial & Retirement Support - Competitive compensation, 401k with employer match, and financial wellness resources
* Time Off & Leave - Paid holidays, flexible vacation time/PSSL, and paid parental leave
* Wellness & Growth - Work life assistance resources, physical wellness perks, mental health support, employee referral program, and BenefitHub for employee discounts
Monogram Health is a leading multispecialty provider of in-home, evidence-based care for the most complex of patients who have multiple chronic conditions. Monogram health takes a comprehensive and personalized approach to a person's health, treating not only a disease, but all of the chronic conditions that are present - such as diabetes, hypertension, chronic kidney disease, heart failure, depression, COPD, and other metabolic disorders.
Monogram Health employs a robust clinical team, leveraging specialists across multiple disciplines including nephrology, cardiology, endocrinology, pulmonology, behavioral health, and palliative care to diagnose and treat health issues; review and prescribe medication; provide guidance, education, and counselling on a patient's healthcare options; as well as assist with daily needs such as access to food, eating healthy, transportation, financial assistance, and more. Monogram Health is available 24 hours a day, 7 days a week, and on holidays, to support and treat patients in their home.
Monogram Health's personalized and innovative treatment model is proven to dramatically improve patient outcomes and quality of life while reducing medical costs across the health care continuum.
$75k-103k yearly est. 60d+ ago
Data Scientist
Signature Science, LLC 4.4
New Jersey jobs
The primary purpose of this position is to serve as the data scientist with a split portfolio between the Atlantic City office and the Austin chemistry group.
Essential Duties and Responsibilities:
Performs data analytics, specifically data clean-up, data processing, predictive modeling, chemometric statistical modeling and analysis, multivariate data analysis, machine learning, and/or data mining, as related to scientific data.
Applies technical skills to plan and execute assigned project work including development of computational models, programming of detection algorithms, and machine learning.
Maintains operational capabilities of computation assets as needed by project requirements.
Leads meetings with company clients by preparing and presenting meeting materials in meetings.
Appropriately annotates project developed computer code through comments and user manuals.
Presents technical results through the drafting of technical reports.
Presents experimental results and recommended actions at internal project meetings.
Supports business development efforts as needed by drafting technical sections of proposals, providing proposal review, assessing levels of effort required to complete proposed work, and brainstorming technical solutions to client problems.
Other duties as assigned.
Required Knowledge, Skills & Abilities:
Ability to plan sequence of experiments to answer complicated technical questions
Ability to lead group of co-workers in execution of a task
Software programming proficiency with Java, C, R, Python, and/or MATLAB
Working knowledge of statistics as it applies to scientific data
Ability to communicate technical information to non-technical audiences
Team player with a positive attitude
Department of Homeland Security Suitability
Department of Defense Secret Clearance
Working knowledge of software development practices including Agile development and Git version control
Sufficient business knowledge to support proposal efforts
Education/Experience:
Incumbent professional should have a Ph.D. or master's degree in a physical science (preferably chemistry), statistics, or data science and significant experience in computer programming, computational modeling, or software development.
Certificates and Licenses:
None
Clearance:
The ability to obtain a Secret clearance and Department of Homeland Security suitability is required for this position.
Supervisory Responsibilities:
The incumbent professional may oversee junior level staff members performing tasks.
Working Conditions/ Equipment:
The incumbent professional is expected to work and/or be available during regular business hours. He/she should also generally be available via e-mail or phone during non-business hours as needed to address critical issues or emergencies. He/she may be required to travel on behalf of the company up to 25%.
The above job description is not intended to be an all-inclusive list of duties and standards of the position. Incumbents will follow any other instructions and perform any other related duties, as assigned by their supervisor.
Powered by ExactHire:160573