Data Engineer jobs at Wikimedia Foundation - 192 jobs
Software Engineer, MediaWiki Interfaces
Wikimedia Foundation 4.7
Data engineer job at Wikimedia Foundation
The Wikimedia Foundation is looking for a Software Engineer to join the MediaWiki Interfaces team. MediaWiki is the open-source platform that powers Wikipedia and its sister projects - impacting hundreds of millions of Wikipedia users across the globe. Our team owns the web API and other integration interfaces of MediaWiki, enabling other teams to create features that allows users to access, edit, and consume knowledge efficiently and reliably. The MediaWiki page has more information on the kind of work the group undertakes.
This role offers a chance to work on large-scale, mission-driven software while deepening your technical expertise and contributing to systems used globally. You'll collaborate with other engineers and product managers to improve performance, reliability and the responsible adoption of our APIs.
Important: We are a fully remote team. Your working hours will need to overlap with UTC-4 to UTC+2 time zones to best accommodate members of your team around the world.
You will be responsible for:
Designing, developing, and maintaining key parts of the MediaWiki stack;
Writing clean, well-documented, and maintainable code;
Participating in code reviews, design discussions, and technical planning;
Investigating and resolving production issues with support from other engineers;
Contributing to the adoption and advocacy of API-first development, and influencing best practices in system design, SLOs/SLAs, and OpenAPI Specification;
Contributing to improving observability (metrics, logging, monitoring) and reliability of our systems.
Collaborating with the Product team and the Wikimedia community to ensure technical solutions meet user needs.
Working with other engineers to improve the development workflow
Qualities we are looking for:
Clear communicator: You can explain your ideas simply and adapt your communication to different audiences;
Outcome-oriented: You prioritize work that delivers value to users and aligns with organizational goals, breaking it into manageable steps to deliver incremental value;
Accountable: You own your work end-to-end and help establish a culture of responsibility and follow-through;
Candid and respectful: You offer feedback directly, constructively, and with care - and invite the same in return. You support growth by challenging ideas, not people;
Collaborative: You enjoy working with others to solve problems and are open to feedback.
Requirements
Commitment to the Foundation's Mission and Values;
Bachelor's + 2 years of related experience or equivalent work experience;
Proficiency in PHP and JavaScript, or a willingness to learn and work with both;
Familiarity with HTTP and web application architecture;
Familiarity with observability and testing practices.
Additionally, we'd love it if you have:
Experience contributing to open-source software;
Familiarity with MediaWiki;
Experience working in a remote, distributed team;
Interest in API-first development or improving developer experience.
Professional experience in large scale web platforms.
About the Wikimedia Foundation
The Wikimedia Foundation is the nonprofit organization that operates Wikipedia and the other Wikimedia free knowledge projects. Our vision is a world in which every single human can freely share in the sum of all knowledge. We believe that everyone has the potential to contribute something to our shared knowledge, and that everyone should be able to access that knowledge freely. We host Wikipedia and the Wikimedia projects, build software experiences for reading, contributing, and sharing Wikimedia content, support the volunteer communities and partners who make Wikimedia possible, and advocate for policies that enable Wikimedia and free knowledge to thrive.
The Wikimedia Foundation is a charitable, not-for-profit organization that relies on donations. We receive donations from millions of individuals around the world, with an average donation of about $15. We also receive donations through institutional grants and gifts. The Wikimedia Foundation is a United States 501(c)(3) tax-exempt organization with offices in San Francisco, California, USA.
As an equal opportunity employer, the Wikimedia Foundation values having a diverse workforce and continuously strives to maintain an inclusive and equitable workplace. We encourage people with a diverse range of backgrounds to apply. We do not discriminate against any person based upon their race, traits historically associated with race, religion, color, national origin, sex, pregnancy or related medical conditions, parental status, sexual orientation, gender identity, gender expression, age, status as a protected veteran, status as an individual with a disability, genetic information, or any other legally protected characteristics.
The Wikimedia Foundation is a remote-first organization with staff members including contractors based 40+ countries*. Salaries at the Wikimedia Foundation are set in a way that is competitive, equitable, and consistent with our values and culture. The anticipated annual pay range of this position for applicants based within the United States is US$92,267 to US$144,201 with multiple individualized factors, including cost of living in the location, being the determinants of the offered pay. For applicants located outside of the US, the pay range will be adjusted to the country of hire. We neither ask for nor take into consideration the salary history of applicants. The compensation for a successful applicant will be based on their skills, experience and location.
*Please note that we are currently able to hire in the following countries: Australia, Austria, Bangladesh, Belgium, Brazil, Canada, Colombia, Costa Rica, Croatia, Czech Republic, Denmark, Egypt, Estonia, Finland, France, Germany, Ghana, Greece, India, Indonesia, Ireland, Israel, Italy, Kenya, Mexico, Netherlands, Nigeria, Peru, Poland, Singapore, South Africa, Spain, Sweden, Switzerland, Uganda, United Kingdom, United States of America and Uruguay. Our non-US employees are hired through a local third party Employer of Record (EOR).
We periodically review this list to streamline to ensure alignment with our hiring requirements.
All applicants can reach out to their recruiter to understand more about the specific pay range for their location during the interview process.
If you are a qualified applicant requiring assistance or an accommodation to complete any step of the application process due to a disability, you may contact us at ************************ or *****************.
More information
U.S. Benefits & Perks
Applicant Privacy Policy
Wikimedia Foundation
What does the Wikimedia Foundation do?
What makes Wikipedia different from social media platforms?
Our Projects
Our Tech Stack
News from across the Wikimedia movement
Wikimedia Blog
Wikimedia 2030
$92.3k-144.2k yearly Auto-Apply 35d ago
Looking for a job?
Let Zippia find it for you.
Quantitative Developer, Investment Data Platform (IDEA) - Hybrid
CFA Institute 4.7
Boston, MA jobs
A leading financial services firm in Boston is seeking a Quantitative Developer to join its Investment DataEngineering & Analytics team. This role focuses on designing and extending a central research data platform using Python and cloud technologies. The ideal candidate will have a strong background in data modeling and a deep interest in investment data. Responsibilities include developing data models and libraries, collaborating with stakeholders, and driving performance improvements. This full-time role offers extensive benefits and competitive compensation.
#J-18808-Ljbffr
$100k-130k yearly est. 2d ago
Principal Data Scientist
Blue Cross Blue Shield of Minnesota 4.8
Eagan, MN jobs
About Blue Cross and Blue Shield of Minnesota
At Blue Cross and Blue Shield of Minnesota, we are committed to paving the way for everyone to achieve their healthiest life. We are looking for dedicated and motivated individuals who share our vision of transforming healthcare. As a Blue Cross associate, you are joining a culture that is built on values of succeeding together, finding a better way, and doing the right thing. If you are ready to make a difference, join us.
The Impact You Will Have
Blue Cross and Blue Shield of MN is hiring a Principal Data Scientist in Eagan, MN. The Principal Data Scientist will lead the design and deployment of advanced AI solutions, leveraging Large Language Models (LLMs) and innovative agentic AI architectures. This individual will develop scalable systems where multiple AI models collaborate to produce insights and enhance efficiency across the organization.
In this role, you will drive the development of cutting-edge AI models and scalable systems that enable collaboration across multiple models to deliver actionable insights. You will design and implement solutions that ensure accuracy, reliability, and performance while incorporating best practices for fairness and transparency. Working closely with cross-functional teams, you will assess potential risks, optimize workflows, and establish governance processes for model monitoring and lifecycle management. Additionally, you will stay current with emerging technologies and regulatory requirements, mentor data scientists and engineers, and champion innovative approaches to integrating LLMs and agentic AI architectures into enterprise applications.
The ideal candidate brings at least seven years of experience in data science, progressing from predictive analytics and machine learning to sophisticated LLM-driven AI applications. You excel at designing solutions that synthesize data, streamline operations, and deliver impactful insights. Success in this role requires strong partnership skills to translate complex business problems into effective data science and machine learning solutions, while maintaining a balance between innovative advancements and responsible governance for safe, scalable AI deployment.
Your Responsibilities
Lead data science projects to design and implement models and experiments from end to end, including data ingestion and preparation, feature engineering, analysis and modeling, model deployment, performance tracking and documentation.
Act as a mentor to junior data scientists around mature data science practices e.g., readable code, thorough documentation, comprehensive experimentation.
Work with autonomy with business partners to convert ambiguous business problems in clear data science/ML specifications - Use contextual business acumen to convert model predictions/results into impactful insights and provide actionable guidance on risks and limitations.
Work hand-in-hand with product managers, dataengineers, and subject matter experts to ship new models, algorithms and improvements continuously and collaboratively into production.
Use combination of machine learning knowledge and contextual business acumen to convert results into impactful insights and provide actionable guidance on risks and limitations of model.
Write narrative documents for model specification and performance analysis to communicate findings and recommendations to teammates, stakeholders and executive leadership.
Required Skills and Experience
7+ years of related professional experience. All relevant experience including work, education, transferable skills, and military experience will be considered.
Hands-on experience in analytics and data science (specific areas of interest include classification/regression, unsupervised learning, time-series/sequence models, NLP, explainability methods, deep learning).
Advanced proficiency in data science tools such as Python, R, Spark, SQL.
Experience implementing predictive algorithms and associated statistical analysis/inference in a data science/ML workflow manipulating both structured and unstructured data.
Strong communication skills and ability to deliver highly technical results to a diverse audience.
Ability to mentor junior data scientists and lead projects from end-to-end.
High school diploma (or equivalency) and legal authorization to work in the U.S.
Preferred Skills and Experience
Advanced degree in Data Science, Machine Learning, or related field.
Proven experience in building and deploying AI/ML models at scale.
Deep understanding of Responsible AI frameworks, bias detection, and explainability techniques.
Strong communication skills to influence stakeholders and drive organizational change.
Proven experience translating complex analyses into actionable insights.
Ability to break down complex or unclear problems into data-driven hypotheses.
Python: End-to-End Data Workflows is strongly preferred
Experience working with large language models (LLMs)
CI/CD and Infrastructure as code - experience with Docker
Author and maintain Github
Node.js experience
React
Terraform
Database skills
Experience with AWS tools: AWS Bedrock is strongly preferred, DynamoDB, S3, Fargate
Proven ability to work cross-functionally with business stakeholders to understand needs, translate between technical and business objectives, and deliver solutions that align with strategic goals.
Ability to craft compelling narratives using visuals and clear written summaries. Prior experience in healthcare or related field is strongly preferred.
Role DesignationHybrid
Anchored in Connection
Our hybrid approach is designed to balance flexibility with meaningful in-person connection and collaboration. We come together in the office two days each week - most teams designate at least one anchor day to ensure team interaction. These in-person moments foster relationships, creativity, and alignment. The rest of the week you are empowered to work remote.
Compensation and Benefits$100,000.00 - $135,000.00 - $170,000.00 Annual
Pay is based on several factors which vary based on position, including skills, ability, and knowledge the selected individual is bringing to the specific job.
We offer a comprehensive benefits package which may include:
Medical, dental, and vision insurance
Life insurance
401k
Paid Time Off (PTO)
Volunteer Paid Time Off (VPTO)
And more
To discover more about what we have to offer, please review our benefits page.
Equal Employment Opportunity Statement
At Blue Cross and Blue Shield of Minnesota, we are committed to paving the way for everyone to achieve their healthiest life. Blue Cross of Minnesota is an Equal Opportunity Employer and maintains an Affirmative Action plan, as required by Minnesota law applicable to state contractors. All qualified applications will receive consideration for employment without regard to, and will not be discriminated against based on any legally protected characteristic.
Individuals with a disability who need a reasonable accommodation in order to apply, please contact us at: **********************************.
Blue Cross and Blue Shield of Minnesota and Blue Plus are nonprofit independent licensees of the Blue Cross and Blue Shield Association.
$100k-135k yearly Auto-Apply 31d ago
Oracle Orbit Analytics and Data Modelling Consultant
Care It Services 4.3
Cleveland, OH jobs
THEY SHOULD MEET THE TOP 3 MUST HAVE SKILL SETS Title - Oracle Orbit Analytics and Data Modelling ConsultantClient - Fujitsu Job - Remote but 10% travel is required to Work Location - Cleveland , OH (hybrid) Top 3 skills required are1. Should have expertise in building Complex physical and logical data models including good knowledge of data modeling concepts. Must have a very good understanding of normal, complex, and circular joins. 2. Very good understanding of Object and data level securities, and roles and responsibilities.3. Should have strong hold building complex Orbit reports using calculation measures and report insights.
Responsibilities: Develop a comprehensive migration plan from Oracle Discoverer to Orbit Analytics. Conduct a detailed analysis of current Oracle Discoverer reports and dashboards. Design and implement solutions to migrate reports, dashboards, and data models to Orbit Analytics. Development includes understanding of requirements, use of existing data models or working with a data modeler, build of report, technical unit testing, and quality assurance testing before being moved to the UAT environment for CRP and UAT testing. Provide technical leadership and guidance throughout the migration process. Ensure data integrity, accuracy, and consistency during the migration. Optimize the performance of the new Orbit Analytics environment. Utilize in-depth knowledge of Oracle R12 ERP and Oracle Fusion Cloud table structures in the migration process. Ensure seamless integration of data from Oracle R12 ERP and Oracle Fusion Cloud into Orbit Analytics. Collaborate with ERP and cloud teams to understand and address data requirements and challenges. Collaborate with business users to understand their reporting needs and ensure they are met post-migration. Conduct training sessions and create documentation to support end-users in the transition to Orbit Analytics. Communicate progress, challenges, and solutions to stakeholders. Develop and execute test plans to ensure all migrated reports and dashboards function as expected. Identify and resolve any issues that arise during the migration process. Conduct post-migration reviews to ensure all objectives are met. Stay up-to-date with the latest features and best practices in Orbit Analytics. Identify opportunities for further optimization and enhancement of the Orbit Analytics environment. Provide ongoing support and maintenance post-migration.
Qualifications :- Bachelor's degree in Computer Science, Information Technology, or a related field. Masters degree preferred. -Proven experience in Oracle Discoverer and Orbit Analytics. -Extensive experience with Oracle R12 ERP and Oracle Fusion Cloud. - At least 5 years of experience in business intelligence and data analytics. -Demonstrated experience in leading migration projects. Technical Skills: - Strong knowledge of SQL, PL/SQL, and database management. -Proficiency in Orbit Analytics, Oracle Discoverer, Oracle R12 ERP, and Oracle Fusion Cloud. -Familiarity with ETL processes and data warehousing concepts. - Experience with data visualization tools and techniques. Soft Skills: -Excellent problem-solving and analytical skills. - Strong communication and interpersonal skills.- Ability to work independently and as part of a team. -Project management skills with the ability to manage multiple priorities.
Who We Are CARE ITS is a certified Woman-owned and operated minority company (certified as WMBE). At CARE ITS, we are the World Class IT Professionals, helping clients achieve their goals. Care ITS was established in 2010. Since then we have successfully executed several projects with our expert team of professionals with more than 20 years of experience each. We are globally operated with our Head Quarters in Plainsboro, NJ, with focused specialization in Salesforce, Guidewire and AWS. We provide expert solutions to our customers in various business domains.
$72k-106k yearly est. Auto-Apply 60d+ ago
Senior Data Engineer - Digital Pathology - Remote
Mayo Clinic 4.8
Rochester, MN jobs
**Why Mayo Clinic** Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans (************************************** - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
**Benefits Highlights**
+ Medical: Multiple plan options.
+ Dental: Delta Dental or reimbursement account for flexible coverage.
+ Vision: Affordable plan with national network.
+ Pre-Tax Savings: HSA and FSAs for eligible expenses.
+ Retirement: Competitive retirement package to secure your future.
**Responsibilities**
**The Digital Biology team is the advanced technology group for Mayo Clinic Digital Pathology. We are seeking a Senior DataEngineer to execute the technical vision for our shared engineering pod. In this role, you will build, deploy, and optimize the scalable, multimodal data pipelines (pathology, -omics, imaging) that feed our biological foundation models and AI Virtual Cells. Working directly with AI pods and bioinformaticians, you will take ownership of data reliability and velocity, transforming complex raw biological information into high-quality training assets used to develop life-changing diagnostic tools.**
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
**Qualifications**
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred
**The preferred candidate will have experience in:**
+ **SQL**
+ **Python**
+ **Google Cloud Dataflow (Apache Beam)**
+ **Google Cloud BigQuery**
**The preferred candidate will also have the GCP Professional DataEngineer Certification**
**Exemption Status**
Exempt
**Compensation Detail**
$138,257.60 - $200,512.00/ year. Education, experience and tenure may be considered along with internal equity when job offers are extended.
**Benefits Eligible**
Yes
**Schedule**
Full Time
**Hours/Pay Period**
80
**Schedule Details**
M-F daytime hours
100% remote role, the employee must live within the US.
**Weekend Schedule**
NA
**International Assignment**
No
**Site Description**
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is. (*****************************************
**Equal Opportunity**
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the "EOE is the Law" (**************************** . Mayo Clinic participates in E-Verify (******************************************************************************************** and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
**Recruiter**
Laura Percival
**Equal opportunity**
As an Affirmative Action and Equal Opportunity Employer Mayo Clinic is committed to creating an inclusive environment that values the diversity of its employees and does not discriminate against any employee or candidate. Women, minorities, veterans, people from the LGBTQ communities and people with disabilities are strongly encouraged to apply to join our teams. Reasonable accommodations to access job openings or to apply for a job are available.
$138.3k-200.5k yearly 5d ago
Senior Data Engineer
Mayo Clinic 4.8
Rochester, MN jobs
**Why Mayo Clinic** Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans (************************************** - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
**Benefits Highlights**
+ Medical: Multiple plan options.
+ Dental: Delta Dental or reimbursement account for flexible coverage.
+ Vision: Affordable plan with national network.
+ Pre-Tax Savings: HSA and FSAs for eligible expenses.
+ Retirement: Competitive retirement package to secure your future.
**Responsibilities**
We are seeking a talented Senior DataEngineer to join our Advanced Data Lake (ADL) team. This is an infrastructure-heavy, hybrid cloud role with Google Cloud Platform (GCP) as a core requirement. You will build and operate enterprise data Lakehouse platforms that support large-scale analytics and digital transformation.
Your responsibilities will include architecting and maintaining automated data pipelines for ingesting, transforming, and integrating complex datasets. You will use DataStream for real-time data movement and Dataflow for processing at scale. Composer/Airflow will be leveraged for seamless scheduling, monitoring, and automation of pipeline operations. Infrastructure provisioning and workflow management will be handled with Terraform and Dataform to ensure reproducibility and adherence to best practices. All code and pipeline assets will be managed through git repositories, with CI/CD automation and streamlined releases enabled by Azure DevOps (ADO). Changes will be governed by ServiceNow processes to ensure traceability, auditability, and operational compliance.
Core duties involve working with cross-functional teams to translate business needs into pipeline specifications, building and optimizing data models for advanced analytics, and maintaining data quality and security throughout all processes. You will automate workflow monitoring and proactively resolve data issues, applying strong technical and problem-solving skills.
The ideal candidate will have proficiency in Python and SQL, with significant experience in Google Cloud Platform (especially Dataflow and DataStream), Terraform, Dataform, and orchestration with Composer/Airflow. Experience managing code in git repositories, working with Azure DevOps workflows, and following ServiceNow change management processes is required. Strong communication skills and the ability to manage multiple priorities in a remote, team-oriented environment are also necessary.
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
This is a full time remote position within the United States. Mayo Clinic will not sponsor or transfer visas for this position including F1 OPT STEM>
**Qualifications**
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred.
Preferred qualifications include hybrid or multi-cloud experience, familiarity with enterprise data governance, metadata, and lineage tools, and experience working in large, regulated environments. A GCP Professional DataEngineer certification is required.
**Exemption Status**
Exempt
**Compensation Detail**
$138,257.60 - $200,512.00 / year
**Benefits Eligible**
Yes
**Schedule**
Full Time
**Hours/Pay Period**
80
**Schedule Details**
Monday - Friday, 8:00 a.m. - 5:00 p.m. May be required to provide 24/7 on-call support.
**Weekend Schedule**
May be required to provide 24/7 on-call support.
**International Assignment**
No
**Site Description**
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is. (*****************************************
**Equal Opportunity**
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the "EOE is the Law" (**************************** . Mayo Clinic participates in E-Verify (******************************************************************************************** and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
**Recruiter**
Ted Keefe
**Equal opportunity**
As an Affirmative Action and Equal Opportunity Employer Mayo Clinic is committed to creating an inclusive environment that values the diversity of its employees and does not discriminate against any employee or candidate. Women, minorities, veterans, people from the LGBTQ communities and people with disabilities are strongly encouraged to apply to join our teams. Reasonable accommodations to access job openings or to apply for a job are available.
$138.3k-200.5k yearly 5d ago
Senior Data Engineer
Mayo Clinic Health System 4.8
Rochester, MN jobs
Why Mayo Clinic Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
Benefits Highlights
* Medical: Multiple plan options.
* Dental: Delta Dental or reimbursement account for flexible coverage.
* Vision: Affordable plan with national network.
* Pre-Tax Savings: HSA and FSAs for eligible expenses.
* Retirement: Competitive retirement package to secure your future.
Responsibilities
We are seeking a talented Senior DataEngineer to join our Advanced Data Lake (ADL) team. This is an infrastructure-heavy, hybrid cloud role with Google Cloud Platform (GCP) as a core requirement. You will build and operate enterprise data Lakehouse platforms that support large-scale analytics and digital transformation.
Your responsibilities will include architecting and maintaining automated data pipelines for ingesting, transforming, and integrating complex datasets. You will use DataStream for real-time data movement and Dataflow for processing at scale. Composer/Airflow will be leveraged for seamless scheduling, monitoring, and automation of pipeline operations. Infrastructure provisioning and workflow management will be handled with Terraform and Dataform to ensure reproducibility and adherence to best practices. All code and pipeline assets will be managed through git repositories, with CI/CD automation and streamlined releases enabled by Azure DevOps (ADO). Changes will be governed by ServiceNow processes to ensure traceability, auditability, and operational compliance.
Core duties involve working with cross-functional teams to translate business needs into pipeline specifications, building and optimizing data models for advanced analytics, and maintaining data quality and security throughout all processes. You will automate workflow monitoring and proactively resolve data issues, applying strong technical and problem-solving skills.
The ideal candidate will have proficiency in Python and SQL, with significant experience in Google Cloud Platform (especially Dataflow and DataStream), Terraform, Dataform, and orchestration with Composer/Airflow. Experience managing code in git repositories, working with Azure DevOps workflows, and following ServiceNow change management processes is required. Strong communication skills and the ability to manage multiple priorities in a remote, team-oriented environment are also necessary.
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
This is a full time remote position within the United States. Mayo Clinic will not sponsor or transfer visas for this position including F1 OPT STEM>
Qualifications
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred.
Preferred qualifications include hybrid or multi-cloud experience, familiarity with enterprise data governance, metadata, and lineage tools, and experience working in large, regulated environments. A GCP Professional DataEngineer certification is required.
Exemption Status
Exempt
Compensation Detail
$138,257.60 - $200,512.00 / year
Benefits Eligible
Yes
Schedule
Full Time
Hours/Pay Period
80
Schedule Details
Monday - Friday, 8:00 a.m. - 5:00 p.m. May be required to provide 24/7 on-call support.
Weekend Schedule
May be required to provide 24/7 on-call support.
International Assignment
No
Site Description
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is.
Equal Opportunity
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the 'EOE is the Law'. Mayo Clinic participates in E-Verify and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
Recruiter
Ted Keefe
$138.3k-200.5k yearly 5d ago
Senior Data Engineer - Digital Pathology - Remote
Mayo Clinic Health System 4.8
Rochester, MN jobs
Why Mayo Clinic Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
Benefits Highlights
* Medical: Multiple plan options.
* Dental: Delta Dental or reimbursement account for flexible coverage.
* Vision: Affordable plan with national network.
* Pre-Tax Savings: HSA and FSAs for eligible expenses.
* Retirement: Competitive retirement package to secure your future.
Responsibilities
The Digital Biology team is the advanced technology group for Mayo Clinic Digital Pathology. We are seeking a Senior DataEngineer to execute the technical vision for our shared engineering pod. In this role, you will build, deploy, and optimize the scalable, multimodal data pipelines (pathology, -omics, imaging) that feed our biological foundation models and AI Virtual Cells. Working directly with AI pods and bioinformaticians, you will take ownership of data reliability and velocity, transforming complex raw biological information into high-quality training assets used to develop life-changing diagnostic tools.
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
Qualifications
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred
The preferred candidate will have experience in:
* SQL
* Python
* Google Cloud Dataflow (Apache Beam)
* Google Cloud BigQuery
The preferred candidate will also have the GCP Professional DataEngineer Certification
Exemption Status
Exempt
Compensation Detail
$138,257.60 - $200,512.00/ year. Education, experience and tenure may be considered along with internal equity when job offers are extended.
Benefits Eligible
Yes
Schedule
Full Time
Hours/Pay Period
80
Schedule Details
M-F daytime hours 100% remote role, the employee must live within the US.
Weekend Schedule
NA
International Assignment
No
Site Description
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is.
Equal Opportunity
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the 'EOE is the Law'. Mayo Clinic participates in E-Verify and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
Recruiter
Laura Percival
$138.3k-200.5k yearly 5d ago
Data Scientist
Cdc Foundation 4.6
Oklahoma jobs
The Data Scientist will play a crucial role in advancing the CDC Foundation's mission by leveraging data to inform strategic decisions and initiatives in a public health organization. This role is aligned to the Workforce Acceleration Initiative (WAI). WAI is a federally funded CDC Foundation program with the goal of helping the nation's public health agencies by providing them with the technology and data experts they need to accelerate their information system improvements.
Working within the Chickasaw Nation Department of Health (CNDH), Division of Public Health, the Data Scientist will utilize advanced analytics, statistical techniques, and machine learning algorithms to derive insights that support public health efforts. The Data Scientist will use training, mentorship, and project-based learning approaches to develop staff capacity for statistical modeling, data visualization, and data communication among the Division of Public Health team, while facilitating needed data governance and tribal data sovereignty frameworks.
The Data Scientist will be hired by the CDC Foundation and assigned to the CNDH Division of Public Health. This position is eligible for a fully remote work arrangement for U.S. based candidates.
Responsibilities
● Develop, implement, and improve data analysis and visualization tools for use by organization staff, to provide timely, relevant information that informs decisions affecting the public's health.
● Analyze diverse datasets related to public health issues to identify trends, patterns, and correlations.
● Apply statistical methods and machine learning algorithms to extract actionable insights.
● Develop predictive models to anticipate disease patterns, assess risk factors, and guide intervention strategies.
● Continuously optimize algorithms for enhanced accuracy and performance.
● Create compelling visualizations and reports to communicate findings to partners and decision-makers.
● Present data-driven insights in a clear and understandable manner to facilitate informed decision-making.
● Collaborate with the public health organization and its partners to understand their data needs and objectives.
● Provide data-driven support and guidance to inform public health policies and initiatives.
● Stay abreast of emerging trends, technologies, and methodologies in data science and analytics.
● Explore innovative approaches to address complex public health challenges and improve data analysis capabilities.
● Up to 10% domestic travel may be required.
Qualifications
● Bachelor's degree or higher in Data Science, Statistics, Epidemiology, or related field. Master's or PhD in related field preferred.
● Minimum 5 years of relevant professional experience
● Proficiency in programming languages such as Python or R.
● Experience with data manipulation and analysis tools (e.g., SQL, Pandas, NumPy).
● Knowledge of machine learning frameworks (e.g., TensorFlow, Scikit-learn).
● Experience with data visualization tools (e.g., Tableau, Power BI).
● Strong analytical thinking and problem-solving abilities.
● Ability to interpret complex datasets and derive meaningful insights.
● Excellent verbal and written communication skills.
● Ability to convey technical concepts to non-technical partners effectively.
● Flexibility to adapt to evolving project requirements and priorities.
● Professional certifications in data science, machine learning, or public health analytics preferred.
● Outstanding interpersonal and teamwork skills; collegial; energetic; and able to develop productive relationships with colleagues, partners, and partners.
● Demonstrated ability to work well independently and within teams.
● Experience working with tribal nations and tribal data sovereignty is preferred.
● Experience working in a virtual environment with remote partners and teams.
● Proficiency in Microsoft Office
Job Highlights
● Location: Remote, must be based in the United States, Up to 10% domestic travel may be required.
● Salary Range: $92,700-$134,275 per year, plus benefits. Individual salary offers will be based on experience and qualifications unique to each candidate.
● Position Type: Grant funded, limited-term opportunity
● Position End Date: June 30, 2027
Special Notes
This role is involved in a dynamic public health program. As such, roles and responsibilities are subject to change as situations evolve. Roles and responsibilities listed above may be expanded upon or updated to match priorities and needs, once written approval is received by the CDC Foundation in order to best support the public health programming.
The CDC Foundation is a smoke-free environment.
Relocation expenses are not included
All qualified applicants will receive consideration for employment and will not be discriminated against on the basis of race, color, religion, sex, national origin, age, mental or physical disabilities, veteran status, and all other characteristics protected by law.
We comply with all applicable laws including E.O. 11246 and the Vietnam Era Readjustment Assistance Act of 1974 governing employment practices and do not discriminate on the basis of any unlawful criteria in accordance with 41 C.F.R. §§ 60-300.5(a)(12) and 60-741.5(a)(7). As a federal government contractor, we take affirmative action on behalf of protected veterans.
About the CDC Foundation
The CDC Foundation helps CDC save and improve lives by unleashing the power of collaboration between CDC, philanthropies, corporations, organizations and individuals to protect the health, safety and security of America and the world. The CDC Foundation is the go-to nonprofit authorized by Congress to mobilize philanthropic partners and private-sector resources to support CDC's critical health protection mission. The CDC Foundation manages hundreds of programs each year impacting a variety of health threats from chronic disease conditions including cardiovascular disease and cancer, to infectious diseases like rotavirus and HIV, to emergency responses, including COVID-19 and Ebola. Visit ********************* for more information.
$92.7k-134.3k yearly Auto-Apply 8d ago
Senior Data Engineer - Digital Pathology - Remote
Mayo Clinic 4.8
Rochester, MN jobs
The Digital Biology team is the advanced technology group for Mayo Clinic Digital Pathology. We are seeking a Senior DataEngineer to execute the technical vision for our shared engineering pod. In this role, you will build, deploy, and optimize the scalable, multimodal data pipelines (pathology, -omics, imaging) that feed our biological foundation models and AI Virtual Cells. Working directly with AI pods and bioinformaticians, you will take ownership of data reliability and velocity, transforming complex raw biological information into high-quality training assets used to develop life-changing diagnostic tools.
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred
The preferred candidate will have experience in:
SQL
Python
Google Cloud Dataflow (Apache Beam)
Google Cloud BigQuery
The preferred candidate will also have the GCP Professional DataEngineer Certification
$89k-144k yearly est. Auto-Apply 6d ago
Senior Data Engineer
Mayo Clinic 4.8
Rochester, MN jobs
We are seeking a talented Senior DataEngineer to join our Advanced Data Lake (ADL) team. This is an infrastructure-heavy, hybrid cloud role with Google Cloud Platform (GCP) as a core requirement. You will build and operate enterprise data Lakehouse platforms that support large-scale analytics and digital transformation.
Your responsibilities will include architecting and maintaining automated data pipelines for ingesting, transforming, and integrating complex datasets. You will use DataStream for real-time data movement and Dataflow for processing at scale. Composer/Airflow will be leveraged for seamless scheduling, monitoring, and automation of pipeline operations. Infrastructure provisioning and workflow management will be handled with Terraform and Dataform to ensure reproducibility and adherence to best practices. All code and pipeline assets will be managed through git repositories, with CI/CD automation and streamlined releases enabled by Azure DevOps (ADO). Changes will be governed by ServiceNow processes to ensure traceability, auditability, and operational compliance.
Core duties involve working with cross-functional teams to translate business needs into pipeline specifications, building and optimizing data models for advanced analytics, and maintaining data quality and security throughout all processes. You will automate workflow monitoring and proactively resolve data issues, applying strong technical and problem-solving skills.
The ideal candidate will have proficiency in Python and SQL, with significant experience in Google Cloud Platform (especially Dataflow and DataStream), Terraform, Dataform, and orchestration with Composer/Airflow. Experience managing code in git repositories, working with Azure DevOps workflows, and following ServiceNow change management processes is required. Strong communication skills and the ability to manage multiple priorities in a remote, team-oriented environment are also necessary.
Develops and deploys data pipelines, integrations and transformations to support analytics and machine learning applications and solutions as part of an assigned product team using various open-source programming languages and vended software to meet the desired design functionality for products and programs. The position requires maintaining an understanding of the organization's current solutions, coding languages, tools, and regularly requires the application of independent judgment. May provide consultative services to departments/divisions and leadership committees. Demonstrated experience in designing, building, and installing data systems and how they are applied to the Department of Data & Analytics technology framework is required. Candidate will partner with product owners and Analytics and Machine Learning delivery teams to identify and retrieve data, conduct exploratory analysis, pipeline and transform data to help identify and visualize trends, build and validate analytical models, and translate qualitative and quantitative assessments into actionable insights.
This is a full time remote position within the United States. Mayo Clinic will not sponsor or transfer visas for this position including F1 OPT STEM>
Qualifications
A Bachelor's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of five years of professional or research experience in data visualization, dataengineering, analytical modeling techniques; OR an Associate's degree in a relevant field such as engineering, mathematics, computer science, information technology, health science, or other analytical/quantitative field and a minimum of seven years of professional or research experience in data visualization, dataengineering, analytical modeling techniques. In-depth business or practice knowledge will also be considered.
Incumbent must have the ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills, time management skills, and demonstrated experience working on cross functional teams are required. Requires strong analytical skills and the ability to identify and recommend solutions and a commitment to customer service. The position requires excellent verbal and written communication skills, attention to detail, and a high capacity for learning and problem resolution.
Advanced experience in SQL is required. Strong Experience in scripting languages such as Python, JavaScript, PHP, C++ or Java & API integration is required. Experience in hybrid data processing methods (batch and streaming) such as Apache Spark, Hive, Pig, Kafka is required. Experience with big data, statistics, and machine learning is required. The ability to navigate linux and windows operating systems is required. Knowledge of workflow scheduling (Apache Airflow Google Composer), Infrastructure as code (Kubernetes, Docker) CI/CD (Jenkins, Github Actions) is preferred. Experience in DataOps/DevOps and agile methodologies is preferred. Experience with hybrid data virtualization such as Denodo is preferred. Working knowledge of Tableau, Power BI, SAS, ThoughtSpot, DASH, d3, React, Snowflake, SSIS, and Google Big Query is preferred.
Google Cloud Platform (GCP) certification is preferred.
Preferred qualifications include hybrid or multi-cloud experience, familiarity with enterprise data governance, metadata, and lineage tools, and experience working in large, regulated environments. A GCP Professional DataEngineer certification is required.
$89k-144k yearly est. Auto-Apply 5d ago
Data Engineer
Mayo Clinic Health System 4.8
Rochester, MN jobs
Why Mayo Clinic Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
Benefits Highlights
* Medical: Multiple plan options.
* Dental: Delta Dental or reimbursement account for flexible coverage.
* Vision: Affordable plan with national network.
* Pre-Tax Savings: HSA and FSAs for eligible expenses.
* Retirement: Competitive retirement package to secure your future.
Responsibilities
The DataEngineer gains access to data across the organization and provides ongoing analysis of the data by monitoring, profiling and analyzing databases. Requires a mix of functional, data and technical skills. The right candidate must be able to understand business requirements, translate them into information needs and implement those requirements using data available. The hire will be responsible for expanding and optimizing data architecture, as well as optimizing data flow and collection for cross functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems. The DataEngineer will support our software developers, database architects, and data analysts and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects.
Core job duties include:
* Assemble large, complex data sets that meet functional / non-functional business requirements.
* Strong knowledge of SQL required. Ability to identify sets and subsets of information across multiple joins or unions of tables is preferred in addition to writing and troubleshooting SQL queries for data mining
* Perform complex data analysis and investigation for customer requests to explain results and to make appropriate recommendations.
* Strong understanding of data modeling concepts
* Problem solver with the initiative to think critically to identify improvement opportunities (error detection, error correction, root cause analysis)
* Understand ETL that will aid in verification and testing of data
* Build processes supporting data transformation, data structures, metadata, dependency and workload management.
* A successful history of manipulating, processing and extracting value from large disconnected datasets.
* Analyze business objectives and develop data solutions to meet customer needs.
* Demonstrated ability to effectively participate in multiple, concurrent projects
* Improve and customize current data solutions to meet business functional and non-functional requirements.
* Research new and existing data sources in order to contribute to new development, improve data management processes, and make recommendations for data quality initiatives.
* Perform periodic data quality reviews for internal and external data.
* Ensure timely resolution of queries and data issues.
* Look for new ways to find and collect data by researching potential new sources of information.
* Work with data and analytics experts to strive for greater functionality in our data systems.
This is a full time, remote position within the United States. Mayo Clinic will not sponsor or transfer visas for this position including F1 OPT STEM.
During the selection process you may participate in an OnDemand (pre-recorded) interview that you can complete at your convenience. During the OnDemand interview, a question will appear on your screen, and you will have time to consider each question before responding. You will have the opportunity to re-record your answer to each question - Mayo Clinic will only see the final recording. The complete interview will be reviewed by a Mayo Clinic staff member and you will be notified of next steps.
A Codility Test may be required as part of the candidate selection process.
Qualifications
Bachelor's degree in Computer Science, Engineering, or related field from an accredited University or College; OR an Associate's degree in Computer Science, Engineering, or related field from an accredited University or College with 2 years of experience. Demonstrated ability to analyze and profile data as a means to address various business problems through leveraging advanced data modeling, source system databases, or data mining techniques, is required. May provide consultative services to departments/divisions and committees. Demonstrated application of several problem-solving methodologies, planning techniques, continuous improvement methods, and analytical tools and methodologies (e.g. data analysis, data profiling, modeling, etc.) required. Incumbent must have ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills and time management skills are required. Requires strong analytical skills and the ability to identify and recommend solutions, advanced computer application skills and a commitment to customer service. Experience with data analysis, quality, and profiling; including data exploration tools including but not limited to Rapid SQL, AQT, Information Analyzer, and Informatics.
Experience with Python, Oracle OCI/FDI/ODI, Automation/AI Experiences, Google Cloud and/or Microsoft Azure dataengineering preferred.
Exemption Status
Exempt
Compensation Detail
$ 100,339.20 - $140,462.40 / year
Benefits Eligible
Yes
Schedule
Full Time
Hours/Pay Period
80
Schedule Details
Monday - Friday, 8am - 5pm. Rotating on-call 24x7 schedule.
Weekend Schedule
As needed for rotating on-call 24x7 schedule.
International Assignment
No
Site Description
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is.
Equal Opportunity
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the 'EOE is the Law'. Mayo Clinic participates in E-Verify and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
Recruiter
Ted Keefe
$100.3k-140.5k yearly 13d ago
Data Engineer
Mayo Clinic 4.8
Rochester, MN jobs
**Why Mayo Clinic** Mayo Clinic is top-ranked in more specialties than any other care provider according to U.S. News & World Report. As we work together to put the needs of the patient first, we are also dedicated to our employees, investing in competitive compensation and comprehensive benefit plans (************************************** - to take care of you and your family, now and in the future. And with continuing education and advancement opportunities at every turn, you can build a long, successful career with Mayo Clinic.
**Benefits Highlights**
+ Medical: Multiple plan options.
+ Dental: Delta Dental or reimbursement account for flexible coverage.
+ Vision: Affordable plan with national network.
+ Pre-Tax Savings: HSA and FSAs for eligible expenses.
+ Retirement: Competitive retirement package to secure your future.
**Responsibilities**
The DataEngineer gains access to data across the organization and provides ongoing analysis of the data by monitoring, profiling and analyzing databases. Requires a mix of functional, data and technical skills. The right candidate must be able to understand business requirements, translate them into information needs and implement those requirements using data available. The hire will be responsible for expanding and optimizing data architecture, as well as optimizing data flow and collection for cross functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems. The DataEngineer will support our software developers, database architects, and data analysts and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects.
Core job duties include:
- Assemble large, complex data sets that meet functional / non-functional business requirements.
- Strong knowledge of SQL required. Ability to identify sets and subsets of information across multiple joins or unions of tables is preferred in addition to writing and troubleshooting SQL queries for data mining
- Perform complex data analysis and investigation for customer requests to explain results and to make appropriate recommendations.
- Strong understanding of data modeling concepts
- Problem solver with the initiative to think critically to identify improvement opportunities (error detection, error correction, root cause analysis)
- Understand ETL that will aid in verification and testing of data
- Build processes supporting data transformation, data structures, metadata, dependency and workload management.
- A successful history of manipulating, processing and extracting value from large disconnected datasets.
- Analyze business objectives and develop data solutions to meet customer needs.
- Demonstrated ability to effectively participate in multiple, concurrent projects
- Improve and customize current data solutions to meet business functional and non-functional requirements.
- Research new and existing data sources in order to contribute to new development, improve data management processes, and make recommendations for data quality initiatives.
- Perform periodic data quality reviews for internal and external data.
- Ensure timely resolution of queries and data issues.
- Look for new ways to find and collect data by researching potential new sources of information.
- Work with data and analytics experts to strive for greater functionality in our data systems.
This is a full time, remote position within the United States. Mayo Clinic will not sponsor or transfer visas for this position including F1 OPT STEM.
During the selection process you may participate in an OnDemand (pre-recorded) interview that you can complete at your convenience. During the OnDemand interview, a question will appear on your screen, and you will have time to consider each question before responding. You will have the opportunity to re-record your answer to each question - Mayo Clinic will only see the final recording. The complete interview will be reviewed by a Mayo Clinic staff member and you will be notified of next steps.
A Codility Test may be required as part of the candidate selection process.
**Qualifications**
Bachelor's degree in Computer Science, Engineering, or related field from an accredited University or College; OR an Associate's degree in Computer Science, Engineering, or related field from an accredited University or College with 2 years of experience. Demonstrated ability to analyze and profile data as a means to address various business problems through leveraging advanced data modeling, source system databases, or data mining techniques, is required. May provide consultative services to departments/divisions and committees. Demonstrated application of several problem-solving methodologies, planning techniques, continuous improvement methods, and analytical tools and methodologies (e.g. data analysis, data profiling, modeling, etc.) required. Incumbent must have ability to manage a varied workload of projects with multiple priorities and stay current on healthcare trends and enterprise changes. Interpersonal skills and time management skills are required. Requires strong analytical skills and the ability to identify and recommend solutions, advanced computer application skills and a commitment to customer service. Experience with data analysis, quality, and profiling; including data exploration tools including but not limited to Rapid SQL, AQT, Information Analyzer, and Informatics.
Experience with Python, Oracle OCI/FDI/ODI, Automation/AI Experiences, Google Cloud and/or Microsoft Azure dataengineering preferred.
**Exemption Status**
Exempt
**Compensation Detail**
$ 100,339.20 - $140,462.40 / year
**Benefits Eligible**
Yes
**Schedule**
Full Time
**Hours/Pay Period**
80
**Schedule Details**
Monday - Friday, 8am - 5pm. Rotating on-call 24x7 schedule.
**Weekend Schedule**
As needed for rotating on-call 24x7 schedule.
**International Assignment**
No
**Site Description**
Just as our reputation has spread beyond our Minnesota roots, so have our locations. Today, our employees are located at our three major campuses in Phoenix/Scottsdale, Arizona, Jacksonville, Florida, Rochester, Minnesota, and at Mayo Clinic Health System campuses throughout Midwestern communities, and at our international locations. Each Mayo Clinic location is a special place where our employees thrive in both their work and personal lives. Learn more about what each unique Mayo Clinic campus has to offer, and where your best fit is. (*****************************************
**Equal Opportunity**
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, gender identity, sexual orientation, national origin, protected veteran status or disability status. Learn more about the "EOE is the Law" (**************************** . Mayo Clinic participates in E-Verify (******************************************************************************************** and may provide the Social Security Administration and, if necessary, the Department of Homeland Security with information from each new employee's Form I-9 to confirm work authorization.
**Recruiter**
Ted Keefe
**Equal opportunity**
As an Affirmative Action and Equal Opportunity Employer Mayo Clinic is committed to creating an inclusive environment that values the diversity of its employees and does not discriminate against any employee or candidate. Women, minorities, veterans, people from the LGBTQ communities and people with disabilities are strongly encouraged to apply to join our teams. Reasonable accommodations to access job openings or to apply for a job are available.
$100.3k-140.5k yearly 13d ago
John Snow Labs US-Based Healthcare Data Scientist
John Snow Labs 4.4
Delaware City, DE jobs
John Snow Labs is an award-winning AI and NLP company, accelerating progress in data science by providing state-of-the-art software, data, and models. Founded in 2015, it helps healthcare and life science companies build, deploy, and operate AI products and services. John Snow Labs is the winner of the 2018 AI Solution Provider of the Year Award, the 2019 AI Platform of the Year Award, the 2019 International Data Science Foundation Technology award, and the 2020 AI Excellence Award.
John Snow Labs is the developer of Spark NLP - the world's most widely used NLP library in the enterprise - and is the world's leading provider of state-of-the-art clinical NLP software, powering some of the world's largest healthcare & pharma companies. John Snow Labs is a global team of specialists, of which 33% hold a Ph.D. or M.D. and 75% hold at least a Master's degree in disciplines covering data science, medicine, software engineering, pharmacy, DevOps and SecOps.
Job Description
John Snow Labs is seeking a highly skilled and motivated Data Scientist to contribute to transformative initiatives within the healthcare industry. The ideal candidate will possess a strong background in developing and optimizing machine learning models, specifically within healthcare contexts. We are looking for a results-oriented individual proficient in training and fine-tuning models, building robust, production-ready model inference pipelines, and conducting comprehensive exploratory data analysis and data enrichment.
Qualifications
Key Responsibilities:
Train, fine tune, and enhance LLM & NLP models using the open-source Python library ecosystem. Experience with LLMs, Generative AI, and deep learning is a significant advantage.
Build data science and dataengineering pipelines specific to analyzing clinical data, such as extracting information from medical text or images, or integrating uncertain information from multiple medical data sources.
Collaborate with our team on customer-facing projects, utilizing your expertise to create advanced machine learning, deep learning, large language models, and time series forecasting pipelines tailored to address specific business needs.
Ensure models are validated for issues like bias, overfitting, and concept drift to ensure reliability and effectiveness.
Engage directly with customers, requiring strong oral and written communication skills to convey complex technical concepts clearly.
Mandatory Skills:
Proven experience in consistently delivering real-world projects covering the key responsibilities. Knowledge that is limited to an academic setting, or to using existing APIs to building applications, is not sufficient for this role.
Hands-on experience with OMOP, FHIR, clinical terminologies, and understanding of the patient journey.
Strong background in healthcare-related fields such as medicine, pharma, bioinformatics, or biostatistics is highly beneficial.
A PhD in a relevant field is preferred but not required if exceptional experience is demonstrated.
Experience with John Snow Labs' technology stack, such as Spark NLP or the medical language models, is a plus.
What We Offer:
A chance to work on cutting-edge problems in healthcare and life sciences, contributing to meaningful projects that impact patient outcomes.
Long-term freelancing contracts with a commitment of at least 30 hours per week. We are seeking individuals, not agencies or teams.
The opportunity to grow your skills and knowledge, working with a team of big data and data science experts in a supportive, collaborative environment.
To apply, please include the words 'John Snow Labs' in your cover letter and detail why you believe you are the best fit for this role. This is more than just a contract - it's a chance to make a real difference.
Additional Information
Our Commitment to You
At John Snow Labs, we believe that diversity is the catalyst of innovation. We're committed to empowering talented people from every background and perspective to thrive.
We are an award-winning global collaborative team focused on helping our customers put artificial intelligence to good use faster. Our website includes The Story of John Snow, and our Social Impact page details how purpose and giving back is part of our DNA. More at JohnSnowLabs.com
We are a fully virtual company, collaborating across 28 countries.
This is a contract opportunity, not a full-time employment role.
This role requires the availability of at least 30-40 hours per week.
$77k-102k yearly est. 60d+ ago
Data Engineer
Cdc Foundation 4.6
Oklahoma jobs
The DataEngineer will play a crucial role in advancing the CDC Foundation's mission by designing, building, and maintaining data infrastructure for a public health organization. This role is aligned to the Workforce Acceleration Initiative (WAI). WAI is a federally funded CDC Foundation program with the goal of helping the nation's public health agencies by providing them with the technology and data experts they need to accelerate their information system improvements.
Working within the Chickasaw Nation Department of Health (CNDH), Division of Public Health, the DataEngineer will collaboratively plan and develop solutions for interoperable data generation, storage, processing, and analysis across data systems relevant to the tribal public health agency. The DataEngineer will collaborate with the CNDH Division of Public Health team and with data content experts, analysts, data scientists, data modelers, warehouse architects, IT staff and other organization staff to design and implement proposed solutions and architectures that meet the needs of the public health agency, all while providing training and mentorship to the Division of Public Health staff for capacity development and system sustainability.
The DataEngineer will be hired by the CDC Foundation and assigned to the CNDH Division of Public Health. This position is eligible for a fully remote work arrangement for U.S. based candidates.
Responsibilities
· Create and manage the systems and pipelines that enable efficient and reliable flow of data, including ingestion, processing, and storage.
· Collect data from various sources, transforming and cleaning it to ensure accuracy and consistency. Load data into storage systems or data warehouses.
· Optimize data pipelines, infrastructure, and workflows for performance and scalability.
· Monitor data pipelines and systems for performance issues, errors, and anomalies, and implement solutions to address them.
· Implement security measures to protect sensitive information.
· Collaborate with data scientists, analysts, and other partners to understand their data needs and requirements, and to ensure that the data infrastructure supports the organization's goals and objectives.
· Collaborate with cross-functional teams to understand data requirements and design scalable solutions that meet business needs.
· Implement and maintain ETL processes to ensure the accuracy, completeness, and consistency of data.
· Design and manage data storage systems, including relational databases, NoSQL databases, and data warehouses.
· Knowledge about industry trends, best practices, and emerging technologies in dataengineering, and incorporating the trends into the organization's data infrastructure.
· Provide technical guidance to other staff.
· Communicate effectively with partners at all levels of the organization to gather requirements, provide updates, and present findings.
· Up to 10% domestic travel may be required.
Qualifications
· Bachelor's degree in Computer Science, Information Technology, Data Science, or a related field.
· Minimum 5 years of relevant professional experience
· Proficiency in programming languages commonly used in dataengineering, such as R, Python, Java, Scala, or SQL. Candidate should be able to implement data automations within existing frameworks as opposed to writing one off scripts.
· Experience with big data technologies and frameworks like Hadoop, Spark, Kafka, and Flink.
· Strong understanding of database systems, including relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra).
· Experience with large/national datasets and Epic Health Record systems is preferred.
· Experience with data visualization tool Power BI.
· Experience regarding engineering best practices such as source control, automated testing, continuous integration and deployment, and peer review.
· Knowledge of data warehousing concepts and tools.
· Experience with cloud computing platforms.
· Expertise in data modeling, ETL (Extract, Transform, Load) processes, and data integration techniques.
· Familiarity with agile development methodologies, software design patterns, and best practices.
· Strong analytical thinking and problem-solving abilities.
· Excellent verbal and written communication skills, including the ability to convey technical concepts to non-technical partners effectively.
· Flexibility to adapt to evolving project requirements and priorities.
· Outstanding interpersonal and teamwork skills; and the ability to develop productive working relationships with colleagues and partners.
· Experience working with tribal nations and tribal data sovereignty is preferred.
· Experience working in a virtual environment with remote partners and teams.
· Proficiency in Microsoft Office.
Job Highlights
· Location: Remote, must be based in the United States, Up to 10% domestic travel may be required.
· Salary Range: $103,500-$143,500 per year, plus benefits. Individual salary offers will be based on experience and qualifications unique to each candidate.
· Position Type: Grant funded, limited-term opportunity
· Position End Date: June 30 2027
Special Notes
This role is involved in a dynamic public health program. As such, roles and responsibilities are subject to change as situations evolve. Roles and responsibilities listed above may be expanded upon or updated to match priorities and needs, once written approval is received by the CDC Foundation in order to best support the public health programming.
The CDC Foundation is a smoke-free environment.
Relocation expenses are not included.
All qualified applicants will receive consideration for employment and will not be discriminated against on the basis of race, color, religion, sex, national origin, age, mental or physical disabilities, veteran status, and all other characteristics protected by law.
We comply with all applicable laws including E.O. 11246 and the Vietnam Era Readjustment Assistance Act of 1974 governing employment practices and do not discriminate on the basis of any unlawful criteria in accordance with 41 C.F.R. §§ 60-300.5(a)(12) and 60-741.5(a)(7). As a federal government contractor, we take affirmative action on behalf of protected veterans.
About the CDC Foundation
The CDC Foundation helps CDC save and improve lives by unleashing the power of collaboration between CDC, philanthropies, corporations, organizations and individuals to protect the health, safety and security of America and the world. The CDC Foundation is the go-to nonprofit authorized by Congress to mobilize philanthropic partners and private-sector resources to support CDC's critical health protection mission. The CDC Foundation manages hundreds of programs each year impacting a variety of health threats from chronic disease conditions including cardiovascular disease and cancer, to infectious diseases like rotavirus and HIV, to emergency responses, including COVID-19 and Ebola. Visit ********************* for more information.
$103.5k-143.5k yearly Auto-Apply 8d ago
Data Engineer
Cdc Foundation 4.6
Illinois jobs
The DataEngineer will play a crucial role in advancing the CDC Foundation's mission by designing, building, and maintaining data infrastructure for a public health organization. This role is aligned to the Workforce Acceleration Initiative (WAI). WAI is a federally funded CDC Foundation program with the goal of helping the nation's public health agencies by providing them with the technology and data experts they need to accelerate their information system improvements.
Working within the Illinois Department of Public Health, to build, optimize, and manage cloud-based data pipelines and ETL processes on the Snowflake platform. This role will focus on implementing high-performance data systems, ensuring the reliability and scalability of our data infrastructure, and enabling robust analytics and insights for the organization. The ideal candidate will have hands-on experience with Snowflake, cloud environments, and dataengineering best practices.
The DataEngineer will be hired by the CDC Foundation and assigned to the Illinois Department of Public Health. This position is eligible for a fully remote work arrangement for U.S. based candidates.
Responsibilities
· Develop a detailed plan for database migration, ETL processes, and data processing applications.
· Design, build, and manage ETL/ELT processes and data pipelines on the Snowflake platform, ensuring the movement of large datasets between various data sources. Develop efficient, scalable data architectures and implement Snowflake best practices, including partitioning, clustering, and query optimization for performance and cost. Collaborate with data scientists, analysts, and Local health departments to integrate diverse data sources into Snowflake, ensuring data is available for analytics and reporting.
· Monitor data pipelines and systems for performance issues, costs, errors and anomalies, and implement solutions to address them.
· Collaborate with the IT Security Team to conduct security and access testing. Implement security measures to protect sensitive information.
· Collaborate with cross-functional teams to understand data requirements and design scalable solutions that meet business needs. Collaborate with Systems Architect on overall system health, focusing on data aspects and data warehouse. Collaborate with Systems Architect on infrastructure assessment, focusing on data aspects.
· Implement and maintain ETL processes to ensure the accuracy, completeness, and consistency of data.
· Design and manage data storage systems, including relational databases, NoSQL databases, and data warehouses.
· Knowledgeable about industry trends, best practices, and emerging technologies in dataengineering, and incorporating the trends into the organization's data infrastructure.
· Provide technical guidance to other staff. Create and maintain clear documentation for ETL processes, data pipelines, data models, and infrastructure setups. Develop training materials and conduct online sessions on accessing and utilizing shared data.
· Communicate effectively with partners at all levels of the organization to gather requirements, provide updates, and present findings.
· Create a data governance framework for secure and compliant data sharing.
· Establish successful connection migration plan for ETL processes and APIs between migrated applications and databases.
· Implement automated processes for data extraction from source systems and loading into the data warehouse.
· Migrate ETL processes and APIs to the cloud environment.
· Up to 10% domestic travel may be required.
Qualifications
· Bachelor's degree in Computer Science, Information Technology, Data Science, or a related field.
· Minimum 5 years of relevant professional experience.
· Proficiency in programming languages commonly used in dataengineering, such as Python, Java, Scala, or SQL. Candidate should be able to implement data automations within existing frameworks as opposed to writing one off scripts.
· Experience with big data technologies and frameworks like Hadoop, Spark, Kafka, and Flink.
· High level of proficiency in Snowflake, including advanced features like Time Travel, Zero-Copy Cloning, and data sharing is required. Experience regarding engineering best practices such as source control, automated testing, continuous integration and deployment, and peer review.
· Knowledge of data warehousing concepts and tools. Familiarity with data lake and lakehouse architectures.
· Expertise in data modeling, ETL (Extract, Transform, Load) processes, and data integration techniques.
· Familiarity with agile development methodologies, software design patterns, and best practices.
· Strong analytical thinking and problem-solving abilities.
· Excellent verbal and written communication skills, including the ability to convey technical concepts to non-technical partners effectively.
· Flexibility to adapt to evolving project requirements and priorities.
· Outstanding interpersonal and teamwork skills; and the ability to develop productive working relationships with colleagues and partners.
· Experience working in a virtual environment with remote partners and teams
· Proficiency in Microsoft Office.
Job Highlights
· Location: Remote, must be based in the United States. Up to 10% domestic travel may be required.
· Salary Range: $103,500-$143,500 per year, plus benefits. Individual salary offers will be based on experience and qualifications unique to each candidate.
· Position Type: Grant funded, limited-term opportunity
· Position End Date: June 30, 2026
· Hours of Work: Individual must routinely be available during core department business hours to collaborate with Illinois Department of Public Health staff.
Special Notes
This role is involved in a dynamic public health program. As such, roles and responsibilities are subject to change as situations evolve. Roles and responsibilities listed above may be expanded upon or updated to match priorities and needs, once written approval is received by the CDC Foundation in order to best support the public health programming.
The CDC Foundation is a smoke-free environment.
Relocation expenses are not included.
All qualified applicants will receive consideration for employment and will not be discriminated against on the basis of race, color, religion, sex, national origin, age, mental or physical disabilities, veteran status, and all other characteristics protected by law.
We comply with all applicable laws including E.O. 11246 and the Vietnam Era Readjustment Assistance Act of 1974 governing employment practices and do not discriminate on the basis of any unlawful criteria in accordance with 41 C.F.R. §§ 60-300.5(a)(12) and 60-741.5(a)(7). As a federal government contractor, we take affirmative action on behalf of protected veterans.
About the CDC Foundation
The CDC Foundation helps CDC save and improve lives by unleashing the power of collaboration between CDC, philanthropies, corporations, organizations and individuals to protect the health, safety and security of America and the world. The CDC Foundation is the go-to nonprofit authorized by Congress to mobilize philanthropic partners and private-sector resources to support CDC's critical health protection mission. The CDC Foundation manages hundreds of programs each year impacting a variety of health threats from chronic disease conditions including cardiovascular disease and cancer, to infectious diseases like rotavirus and HIV, to emergency responses, including COVID-19 and Ebola. Visit ********************* for more information.
$103.5k-143.5k yearly Auto-Apply 6d ago
Data Engineer
Battelle 4.7
Columbus, OH jobs
Battelle delivers when others can't. We conduct research and development, manage national laboratories, design and manufacture products and deliver critical services for our clients-whether they are a multi-national corporation, a small start-up or a government agency.
We recognize and appreciate the value and contributions of individuals from a wide range of backgrounds and experiences and welcome all qualified individuals to apply.
Job Summary
As a DataEngineer within Battelle's Cyber business area, you will support a program mission to deliver comprehensive technical expertise and testing support to our government customer, contributing to the evaluation, customization, and operationalization of cutting-edge systems and payloads ensuring they meet the customer's evolving requirements for effectiveness, safety, and security.
Responsibilities
Design and implement data acquisition systems for test environments, enabling real-time collection, processing, and analysis of technical and operational data from flight tests, component assessments, and system integration activities.
Develop and maintain robust data pipelines that support the ingestion, transformation, and storage of test data from diverse instrumentation sources, ensuring high data integrity and traceability throughout the test lifecycle.
Collaborate with systems engineers and test specialists to define data requirements, integrate instrumentation outputs, and optimize data flows for real-time monitoring and post-test analysis.
Implement real-time data processing solutions to support rapid assessment of system performance, technical maturity, and operational readiness during test events.
Develop and deploy data visualization tools and dashboards that provide stakeholders with actionable insights from test data, including key findings, trends, and risk factors relevant to system selection and integration.
Ensure secure management and compliance of test data, including configuration management, audit readiness, and integration with digital engineering environments (e.g., MBSE repositories, digital twin libraries).
Support the creation of technical analysis reports and comparative matrices by integrating real-time and post-test data from instrumentation, modeling, and simulation outputs.
Contribute to supply chain risk analysis by managing and analyzing hardware/software bill of materials (BOM) data, supporting vulnerability assessments, and maintaining requirements traceability within digital engineering workflows.
Prepare and deliver comprehensive data packages, analysis summaries, and supporting documentation for technical reviews, risk assessments, and operational recommendations.
Key Qualifications
Bachelor's degree in a related field with 5+ years of experience, or an equivalent combination of education and experience.
Strong background in modeling and simulation of complex systems.
Proven ability to lead technical discussions and communicate complex concepts to both technical and non-technical stakeholders.
Strong problem-solving, analytical, and teamwork skills.
Nice to Have
Experience with model-based systems engineering (MBSE) methodologies.
Familiarity with SysML diagramming tools (e.g., Cameo Systems Modeler, IBM Rhapsody).
Experience with DoD Architecture Framework (DoDAF) and/or Unified Architectural Framework (UAF).
Experience with embedded systems, microelectronics, FPGA, or circuit card design.
Programming experience in MATLAB, Python, C++, or other object-oriented languages.
Experience supporting DoD programs.
Active Top-Secret Clearance.
Benefits: Live an Extraordinary Life
We care about your well-being, not just on the job. Battelle offers comprehensive and competitive benefits to help you live your best life.
Balance life through a compressed work schedule: Most of our team follows a flexible, compressed work schedule that allows for every other Friday off-giving you a dedicated day to accomplish things in your personal life without using vacation time.
Enjoy enhanced work flexibility, including a hybrid arrangement: You have options for where and when you work. Our Together with Flexibility model allows you to work 60% in-office and 40% remote, with Monday and Tuesday as common in-office days, dependent on team and position needs.
Take time to recharge: You get paid time off to support work-life balance and keep motivated.
Prioritize wellness: Stay healthy with medical, dental, and vision coverage with wellness incentives and benefits plus a variety of optional supplemental benefits.
Better together: Coverage for partners, gender-affirming care and health support, and family formation support.
Build your financial future: Build financial stability with an industry-leading 401(k) retirement savings plan. For most employees, we put in 5 percent whether you contribute or not, and match your contributions on top of that.
Advance your education: Tuition assistance is available to pursue higher education.
A Work Environment Where You Succeed
For brilliant minds in science, technology, engineering and business operations, Battelle is the place to do the greatest good by solving humanity's most pressing challenges and creating a safer, healthier and more secure world.
You will have the opportunity to thrive in a culture that inspires you to:
Apply your talent to challenging and meaningful projects
Receive select funding to pursue ideas in scientific and technological discovery
Partner with world-class experts in a collaborative environment
Nurture and develop the next generation of scientific leaders
Give back to and improve our communities
Vaccinations & Safety Protocols
Battelle may require employees, based on job duties, work location, and/or its clients' requirements to follow certain safety protocols and to be vaccinated against a variety of viruses, bacteria, and diseases as a condition of employment and continued employment and to provide documentation that they are fully vaccinated. If applicable, Battelle will provide reasonable accommodations based on a qualified disability or medical condition through the Americans with Disabilities Act or the Rehabilitation Act or for a sincerely held religious belief under Title VII of the Civil Rights Act of 1964 (and related state laws).
Battelle is an equal opportunity employer. We provide employment and opportunities for advancement, compensation, training, and growth according to individual merit, without regard to race, color, religion, sex (including pregnancy), national origin, sexual orientation, gender identity or expression, marital status, age, genetic information, disability, veteran-status veteran or military status, or any other characteristic protected under applicable Federal, state, or local law. Our goal is for each staff member to have the opportunity to grow to the limits of their abilities and to achieve personal and organizational objectives. We will support positive programs for equal treatment of all staff and full utilization of all qualified employees at all levels within Battelle.
The above statements are intended to describe the nature and level of work being performed by people assigned to this job. They are not intended to be an exhaustive list of all responsibilities, activities and skills required of staff members. No statement herein is intended to imply any authorities to commit Battelle unless special written permission is granted by Battelle's Legal Department.
For more information about our other openings, please visit ************************
$84k-108k yearly est. 60d+ ago
Senior GCP Data Engineer
Care It Services 4.3
Parsippany-Troy Hills, NJ jobs
Senior GCP DataEngineer Job Title : Senior GCP DataEngineer Location : Parsippany, NJ (Hybrid) Duration : Long Term Contract Experience : 8-10+ years Job Description:We are seeking an experienced and highly skilled Senior GCP DataEngineer to join our team on a long-term contract basis. The ideal candidate will have extensive experience with Google Cloud Platform (GCP) and a strong background in dataengineering. This role is hybrid, requiring a presence in our Parsippany, NJ office.Key Responsibilities:
Design, develop, and maintain data processing systems using GCP tools such as Data fusion, BigQuery, and Cloud Composer.
Write efficient, reusable, and reliable Python code for data processing.
Develop complex SQL queries for data analysis and transformation.
Utilize version control systems like GIT for code management.
Implement and manage workflows using Apache Airflow.
Develop and maintain infrastructure as code using Terraform.
Collaborate with cross-functional teams to gather and analyze requirements.
Ensure data quality and integrity across multiple systems.
Troubleshoot and resolve data processing issues.
Required Skills and Qualifications:
10+ years of experience in dataengineering or related fields.
Extensive experience with GCP tools, including Data fusion, BigQuery, Cloud Composer, and other related services.
Proficiency in Python programming.
Advanced SQL skills.
Experience with version control systems, particularly GIT.
Strong understanding of Apache technologies and GitHub.
Experience with Terraform for infrastructure as code.
Solid experience with Apache Airflow for workflow orchestration.
Excellent problem-solving skills and attention to detail.
Strong communication and collaboration skills.
If you are a motivated and skilled GCP DataEngineer with passion, we invite you to apply at sudheer(@)careits (.) com for this exciting opportunity.
Flexible work from home options available.
Who We Are CARE ITS is a certified Woman-owned and operated minority company (certified as WMBE). At CARE ITS, we are the World Class IT Professionals, helping clients achieve their goals. Care ITS was established in 2010. Since then we have successfully executed several projects with our expert team of professionals with more than 20 years of experience each. We are globally operated with our Head Quarters in Plainsboro, NJ, with focused specialization in Salesforce, Guidewire and AWS. We provide expert solutions to our customers in various business domains.
$88k-129k yearly est. Auto-Apply 60d+ ago
Cloud Data Engineer
Summa Health 4.8
Akron, OH jobs
Summa Health Systems Full Time, Benefit Eligible 1077 Gorge Blvd Akron OH 44310 * remote opportunity* Designs and implements cloud solutions per regulated Healthcare compliance requirements, ensuring the smooth functioning of cloud systems with minimal downtime or service disruptions. Serves as part of the Enterprise Data Services team where design elements and risk remediation are addressed. Consults with the finance, clinical and operational business teams for system and service expectations. Evaluates and selects supportable platforms, establish best practices for cloud operations, and work with IT teams to monitor cloud privacy and address technical issues. Stays updated on cloud computing technologies, compliance, and trends, and effectively communicates operationalizing cloud operations and deployments. Manages life cycles of private cloud technologies (e.g., VMWare).
Formal Education Required:
a. Bachelor's Degree or equivalent in Software Engineering, Computer Science, IT, Mathematics, or Economics or equivalent combination of education and/or experience.
Experience & Training Required:
a. Three (3) years in an IT role (solution or technical cloud architect role using service and hosting solutions such as private/public cloud IaaS, PaaS, and SaaS platforms).
b. Three (3) years' experience with the concepts, philosophies, and tools behind the design of creating a well-informed holistic cloud strategy and managing the adaptation process.
c. Exhibits strong analytical aptitude across various systems and solutions, internal and external to the organization. Analytical thinker with ability to extend knowledge of information systems and applications to business processes and to cross team collaboration. Ability to explain complex, technical concepts to lay audiences and non-IT staff.
d. Evaluating cloud applications, platforms, hardware, and software
e. Developing and organizing cloud systems
f. Collaborate throughout IT&S, the Business, Finance and Summa Health Leadership
Other Skills, Competencies and Qualifications:
a. Understands cloud architecture, data management and deployment methodologies in alignment with customer driven new cloud initiatives.
b. Ability to identify opportunities for automation and mentorship of the support team.
c. Working knowledge of cloud infrastructure to engineer and support public, private and hybrid clouds.
d. Proficiency in a variety of cloud hosts including Amazon Web Services and Microsoft Azure at a minimum.
e. Experience with and a complete understanding of the technology required for cloud computing
f. Experience with the skills to migrate applications to the cloud
g. Be able to build relationships between the customers and the team members
h. The ability and the knowledge to integrate services and tools to any and all areas to be used such as networks and data processing
i. Experience with meeting or exceeding the requirements of cloud provider service-level agreements Experience with AWS and Azure
j. Knowledge of a multitude of programming languages including SQL, Python, R, Java, Node and Object-Oriented Design Principles required.
Level of Physical Demands:
a. Minimal, may occasionally move computer equipment (desktop, laptop, monitor, printer, and peripherals) when necessary.
Equal Opportunity Employer/Veterans/Disabled
Summa Health System is recognized as one of the region's top employers by a number of third party organizations, including NorthCoast 99. Exceptional candidates gravitate to Summa because of its culture, passion for delivering excellent service to our patients and families commitment to our philosophy of servant leadership, collegial working relationships at every level of the organization and competitive pay and benefits.
$45.26/hr - $67.89/hr
The salary range on this job posting/advertising is base salary exclusive of any bonuses or differentials. Many factors, such as years of relevant experience and geographical location are considered when determining the starting rate of pay. We believe in the importance of pay equity and consider internal equity of our current team members when determining offers. Please keep in mind that the range that is listed is the full base salary range. Hiring at the maximum of the range would not be typical.
Summa Health offers a competitive and comprehensive benefits program to include medical, dental, vision, life, paid time off as well as many other benefits.
* Basic Life and Accidental Death & Dismemberment (AD&D)
* Supplemental Life and AD&D
* Dependent Life Insurance
* Short-Term and Long-Term Disability
* Accident Insurance, Hospital Indemnity, and Critical Illness
* Retirement Savings Plan
* Flexible Spending Accounts - Healthcare and Dependent Care
* Employee Assistance Program (EAP)
* Identity Theft Protection
* Pet Insurance
* Education Assistance
* Daily Pay
#LI-AW1
$45.3-67.9 hourly 7d ago
Senior Software Engineer (Security)
Wikimedia Foundation 4.7
Data engineer job at Wikimedia Foundation
Timezone preference: GMT-5 through GMT+1
The Wikimedia Foundation is looking for a Senior Software Engineer to join the Product Safety and Integrity team and build new security features to protect Wikipedia and our other projects. This is a very hands-on engineering role, working alongside a team of engineers and product managers to design and implement new features to protect and reassure our users and to ensure the platform remains resilient against attacks.
You are a seasoned software engineer with experience building security features in large-scale systems. You understand the importance of testing and documentation, and common pitfalls in developing secure web applications. You have a passion for the WMF mission. We operate in a highly transparent manner, and the work we do touches thousands of editors every day.
You will be working primarily on our MediaWiki platform which powers Wikipedia. As a top 10 website, we must meet stringent performance standards while addressing new security challenges such as supporting modern authentication technologies, detecting and preventing platform abuse from bots, and planning and rolling out improvements to our security architecture by defending against emerging security threats.
You are responsible for:
Helping design, develop, and deliver security features, with safety and security in mind
Working with other engineering teams to ensure that they make safe architectural and implementation choices
Leading by example in code review, decision-making, and team culture - fostering transparency, empathy, and collaboration
Developing, reviewing, and deploying security features developed by the Foundation and community members
Performing security maintenance and addressing technical debt in security-critical components
Providing support for application security incidents and operations
Skills and Experience:
We value the right mindset and potential over a checklist of specific experiences, and these are the traits we've identified that make great additions to our team so far.
5+ years of experience as a software engineer, ideally with focus on security
Ability to work effectively in a modern PHP and JavaScript codebase
Driving technical quality and operational excellence by defining and reinforcing standards in testing, observability, and system reliability;
Comfortably and autonomously creating proofs of concept, writing design documents, and breaking down complex projects into actionable tasks to support less experienced team members
Experience in developing secure software or security-related product features
A strong interest in working with a talented security team and learning more specialist security skills such as exploiting and mitigating application-level vulnerabilities
Ability to explain complex security issues and their implications on privacy and risk to non-technical audiences
Sensitivity to the security challenges faced by participants in a large, international project
Experience working in a remote, distributed team
Additionally, we'd love it if you have:
Experience working on anti-abuse mechanisms such as CAPTCHA and bot detection
Previous experience building security countermeasures against attacks on technologies at the web, backend, and database level
Experience finding and fixing security bugs and reviewing code for security gaps
A working knowledge of threat modeling and secure design patterns
Prior experience with MediaWiki or Wikimedia projects
Contributions to open-source software
About the Wikimedia Foundation
The Wikimedia Foundation is the nonprofit organization that operates Wikipedia and the other Wikimedia free knowledge projects. Our vision is a world in which every single human can freely share in the sum of all knowledge. We believe that everyone has the potential to contribute something to our shared knowledge, and that everyone should be able to access that knowledge freely. We host Wikipedia and the Wikimedia projects, build software experiences for reading, contributing, and sharing Wikimedia content, support the volunteer communities and partners who make Wikimedia possible, and advocate for policies that enable Wikimedia and free knowledge to thrive.
The Wikimedia Foundation is a charitable, not-for-profit organization that relies on donations. We receive donations from millions of individuals around the world, with an average donation of about $15. We also receive donations through institutional grants and gifts. The Wikimedia Foundation is a United States 501(c)(3) tax-exempt organization with offices in San Francisco, California, USA.
As an equal opportunity employer, the Wikimedia Foundation values having a diverse workforce and continuously strives to maintain an inclusive and equitable workplace. We encourage people with a diverse range of backgrounds to apply. We do not discriminate against any person based upon their race, traits historically associated with race, religion, color, national origin, sex, pregnancy or related medical conditions, parental status, sexual orientation, gender identity, gender expression, age, status as a protected veteran, status as an individual with a disability, genetic information, or any other legally protected characteristics.
The Wikimedia Foundation is a remote-first organization with staff members including contractors based 40+ countries*. Salaries at the Wikimedia Foundation are set in a way that is competitive, equitable, and consistent with our values and culture. The anticipated annual pay range of this position for applicants based within the United States is US$113,082 to US$175,725 with multiple individualized factors, including cost of living in the location, being the determinants of the offered pay. For applicants located outside of the US, the pay range will be adjusted to the country of hire. We neither ask for nor take into consideration the salary history of applicants. The compensation for a successful applicant will be based on their skills, experience and location.
*Please note that we are currently able to hire in the following countries: Australia, Austria, Bangladesh, Belgium, Brazil, Canada, Colombia, Costa Rica, Croatia, Czech Republic, Denmark, Egypt, Estonia, Finland, France, Germany, Ghana, Greece, India, Indonesia, Ireland, Israel, Italy, Kenya, Mexico, Netherlands, Nigeria, Peru, Poland, Singapore, South Africa, Spain, Sweden, Switzerland, Uganda, United Kingdom, United States of America and Uruguay. Our non-US employees are hired through a local third party Employer of Record (EOR).
We periodically review this list to streamline to ensure alignment with our hiring requirements.
All applicants can reach out to their recruiter to understand more about the specific pay range for their location during the interview process.
If you are a qualified applicant requiring assistance or an accommodation to complete any step of the application process due to a disability, you may contact us at ************************ or *****************.
More information
U.S. Benefits & Perks
Applicant Privacy Policy
Wikimedia Foundation
What does the Wikimedia Foundation do?
What makes Wikipedia different from social media platforms?
Our Projects
Our Tech Stack
News from across the Wikimedia movement
Wikimedia Blog
Wikimedia 2030