Data Warehouse DataEngineer to build, enhance and optimize enterprise analytics. Will design and develop ETL/ELT pirpelines, Lakehouse/Warehouse models and curated datasets. Work closely with BI analysts, BI developers, architects and stakeholders.
Dataengineering and pipeline development
Data Modeling and optimization
Data quality, Governance and security
CI/CD
Azire DevP[s
Lakehouse
OneLake
Microsoft Fabric
$97k-129k yearly est. 4d ago
Looking for a job?
Let Zippia find it for you.
Lead HPC Architect Cybersecurity - High Performance & Computational Data Ecosystem
Icahn School of Medicine at Mount Sinai 4.8
Data engineer job in New York, NY
The Scientific Computing and Data group at the Icahn School of Medicine at Mount Sinai partners with scientists to accelerate scientific discovery. To achieve these aims, we support a cutting-edge high-performance computing and data ecosystem along with MD/PhD-level support for researchers. The group is composed of a high-performance computing team, a clinical data warehouse team and a data services team.
The Lead HPC Architect, Cybersecurity, High Performance Computational and Data Ecosystem, is responsible for designing, implementing, and managing the cybersecurity infrastructure and technical operations of Scientific Computing's computational and data science ecosystem. This ecosystem includes a 25,000+ core and 40+ petabyte usable high-performance computing (HPC) systems, clinical research databases, and a software development infrastructure for local and national projects. The HPC system is the fastest in the world at any academic biomedical center (Top 500 list).
To meet Sinai's scientific and clinical goals, the Lead brings a strategic, tactical and customer-focused vision to evolve the ecosystem to be continually more resilient, secure, scalable and productive for basic and translational biomedical research. The Lead combines deep technical expertise in cybersecurity, HPC systems, storage, networking, and software infrastructure with a strong focus on service, collaboration, and strategic planning for researchers and clinicians throughout the organization and beyond. The Lead is an expert troubleshooter, productive partner and leader of projects. The lead will work with stakeholders to make sure the HPC infrastructure is in compliance with governmental funding agency requirements and to promote efficient resource utilizations for researchers
This position reports to the Director for HPC and Data Ecosystem in Scientific Computing and Data.
Key Responsibilities:
HPC Cybersecurity & System Administration:
Design, implement, and manage all cybersecurity operations within the HPC environment, ensuring alignment with industry standards (NIST, ISO, GDPR, HIPAA, CMMC, NYC Cyber Command, etc.).
Implement best practices for data security, including but not limited to encryption (at rest, in transit, and in use), audit logging, access control, authentication control, configuration managements, secure enclaves, and confidential computing.
Perform full-spectrum HPC system administration: installation, monitoring, maintenance, usage reporting, troubleshooting, backup and performance tuning across HPC applications, web service, database, job scheduler, networking, storage, computes, and hardware to optimize workload efficiency.
Lead resolution of complex cybersecurity and system issues; provide mentorship and technical guidance to team members.
Ensure that all designs and implementations meet cybersecurity, performance, scalability, and reliability goals. Ensure that the design and operation of the HPC ecosystem is productive for research.
Lead the integration of HPC resources with laboratory equipment for data ingestion aligned with all regulatory such as genomic sequencers, microscopy, clinical system etc.
Develop, review and maintain security policies, risk assessments, and compliance documentation accurately and efficiently.
Collaborate with institutional IT, compliance, and research teams to ensure all regulatory, Sinai Policy and operational alignment.
Design and implement hybrid and cloud-integrated HPC solutions using on-premise and public cloud resources.
Partner with other peers regionally, nationally and internationally to discover, propose and deploy a world-class research infrastructure for Mount Sinai.
Stay current with emerging HPC, cloud, and cybersecurity technologies to keep the organization's infrastructure up-to-date.
Work collaboratively, effectively and productively with other team members within the group and across Mount Sinai.
Provide after-hours support as needed.
Perform other duties as assigned or requested.
Requirements:
Bachelor's degree in computer science, engineering or another scientific field. Master's or PhD preferred.
10 years of progressive HPC system administration experience with Enterprise Linux releases including RedHat/CentOS/Rocky Systems, and batch cluster environment.
Experience with all aspects of high-throughput HPC including schedulers (LSF or Slurm), networking (Infiniband/Gigabit Ethernet), parallel file systems and storage, configuration management systems (xCAT, Puppet and/or Ansible), etc.
Proficient in cybersecurity processes, posture, regulations, approaches, protocols, firewalls, data protection in a regulated environment (e.g. finance, healthcare).
In-depth knowledge HIPAA, NIST, FISMA, GDPR and related compliance standards, with prove experience building and maintaining compliant HPC system
Experience with secure enclaves and confidential computing.
Proven ability to provide mentorship and technical leadership to team members.
Proven ability to lead complex projects to completion in collaborative, interdisciplinary settings with minimum guidance.
Excellent analytical ability and troubleshooting skills.
Excellent communication, documentation, collaboration and interpersonal skills. Must be a team player and customer focused.
Scripting and programming experience.
Preferred Experience
Proficient with cloud services, orchestration tools, openshift/Kubernetes cost optimization and hybrid HPC architectures.
Experience with Azure, AWS or Google cloud services.
Experience with LSF job scheduler and GPFS Spectrum Scale.
Experience in a healthcare environment.
Experience in a research environment is highly preferred.
Experience with software that enables privacy-preserving linking of PHI.
Experience with Globus data transfer.
Experience with Web service, SAP HANA, Oracle, SQL, MariaDB and other database technologies.
Strength through Unity and Inclusion
The Mount Sinai Health System is committed to fostering an environment where everyone can contribute to excellence. We share a common dedication to delivering outstanding patient care. When you join us, you become part of Mount Sinai's unparalleled legacy of achievement, education, and innovation as we work together to transform healthcare. We encourage all team members to actively participate in creating a culture that ensures fair access to opportunities, promotes inclusive practices, and supports the success of every individual.
At Mount Sinai, our leaders are committed to fostering a workplace where all employees feel valued, respected, and empowered to grow. We strive to create an environment where collaboration, fairness, and continuous learning drive positive change, improving the well-being of our staff, patients, and organization. Our leaders are expected to challenge outdated practices, promote a culture of respect, and work toward meaningful improvements that enhance patient care and workplace experiences. We are dedicated to building a supportive and welcoming environment where everyone has the opportunity to thrive and advance professionally. Explore this opportunity and be part of the next chapter in our history.
About the Mount Sinai Health System:
Mount Sinai Health System is one of the largest academic medical systems in the New York metro area, with more than 48,000 employees working across eight hospitals, more than 400 outpatient practices, more than 300 labs, a school of nursing, and a leading school of medicine and graduate education. Mount Sinai advances health for all people, everywhere, by taking on the most complex health care challenges of our time - discovering and applying new scientific learning and knowledge; developing safer, more effective treatments; educating the next generation of medical leaders and innovators; and supporting local communities by delivering high-quality care to all who need it. Through the integration of its hospitals, labs, and schools, Mount Sinai offers comprehensive health care solutions from birth through geriatrics, leveraging innovative approaches such as artificial intelligence and informatics while keeping patients' medical and emotional needs at the center of all treatment. The Health System includes more than 9,000 primary and specialty care physicians; 13 joint-venture outpatient surgery centers throughout the five boroughs of New York City, Westchester, Long Island, and Florida; and more than 30 affiliated community health centers. We are consistently ranked by U.S. News & World Report's Best Hospitals, receiving high "Honor Roll" status.
Equal Opportunity Employer
The Mount Sinai Health System is an equal opportunity employer, complying with all applicable federal civil rights laws. We do not discriminate, exclude, or treat individuals differently based on race, color, national origin, age, religion, disability, sex, sexual orientation, gender, veteran status, or any other characteristic protected by law. We are deeply committed to fostering an environment where all faculty, staff, students, trainees, patients, visitors, and the communities we serve feel respected and supported. Our goal is to create a healthcare and learning institution that actively works to remove barriers, address challenges, and promote fairness in all aspects of our organization.
$89k-116k yearly est. 3d ago
Senior Architect - NYC Code, Development & CA Specialist
The Highrise Group
Data engineer job in New York, NY
Highrise | Brooklyn, NY (Hybrid)
Highrise is a full-service Architecture, Expediting, and Development firm focused on complex New York City projects. We are seeking a high-level Senior Architect to join our in-house Brooklyn team and lead projects from design through construction.
This is a hybrid position with 2-3 days remote and the remainder in our Brooklyn office.
Role & Responsibilities
Lead architectural design and documentation for new buildings and major alterations
Manage Construction Administration (CA), including:
RFIs, submittals, shop drawings, and field conditions
Site visits and coordination with GC, consultants, and ownership
Issue resolution during construction
Ensure compliance with NYC Zoning, Building Code, Energy Code, and ADA
Coordinate DOB filings and agency reviews
Collaborate closely with ownership, expeditors, engineers, and development teams
Review work and mentor junior staff
Qualifications
Licensed Architect preferred (NY license a strong plus)
Proven experience in Construction Administration for NYC projects
Deep working knowledge of NYC Code, Zoning, Energy, and ADA
Strong development-driven project background
Advanced proficiency in AutoCAD and Revit
Able to independently manage projects from concept through CO
What We Offer
Very competitive compensation
Generous PTO
Hybrid schedule (2-3 days remote)
Direct access to ownership and decision-making
Long-term growth within a fast-growing firm
High-quality NYC development projects
📍 Location: Brooklyn, NY
🗓 Schedule: Full-Time, Hybrid (2-3 days remote)
$99k-129k yearly est. 5d ago
ETL Talend MDM Architect
Trg 4.6
Data engineer job in New York, NY
Responsibilities: • Develop and test Extract, Transformation, and Loading (ETL) modules based on design specifications • Develop and test ETL Mappings in Talend • Plan, test, and deploy ETL mappings, and database code as part of application build process across the enterprise
• Provide effective communications with all levels of internal and external customers and staff
• Must demonstrate knowledge in the following areas:
o Data Integration
o Data Architecture
o Team Lead experience is a plus
• Understand, analyze, assess and recommend ETL environment from technology strategy and operational standpoint
• Understand and assess source system data issues and recommend solution from data integration standpoint
• Create high level, low level technical design documents for data integration
• Design exceptions handling, audit and data resolution processes
• Performance tune ETL environment
• Conduct proof of concepts
• Estimation of work based on functional requirements documents
• Identify system deficiencies and recommending solutions
• Designing, coding, and writing unit test cases from functional requirements
• Delivering efficient and bug-free ETL packages and documentation
• Maintenance and support of enterprise ETL jobs
• Experience with Talend Hadoop tools is a plus
Basic Qualifications:
• 3+ years of development experience on Talend ETL tools
• 7+ years working with one or more of the following ETL Tools: Talend, Informatica, Ab Initio or Data Stage
• 7+ years proficient experience as a developer
• Bachelor's Degree in Computer Science or equivalent
• Database (Oracle, SQL Server, DB2)
• Database Programming (Complex SQL, PL/SQL development knowledge)
• Data Modeling
• Business Analysis
• Top level performer with ability to work independently in short time frames
• Proficient working in a Linux environment
• Experience in scripting languages (Shell, Python or Perl)
• 5+ years of experience deploying large scale projects ETL projects that
• 3+ years of experience in a development lead position
• Data analysis, data mapping, data loading, and data validation
• Understand reusability, parameterization, workflow design, etc.
• Thorough understanding of Entire life cycle of Software and various Software Engineering Methodologies
• Performance tuning of interfaces that extract, transform and load tens of millions of records
• Knowledge of Hadoop ecosystem technologies is a plus
Additional Information
If you are comfortable with the position and location then please revert me back at the earliest with your updated resume and following details or I would really appreciate if you can call me back on my number.
Full Name:
Email:
Skype id:
Contact Nos.:
Current Location:
Open to relocate:
Start Availability:
Work Permit:
Flexible time for INTERVIEW:
Current Company:
Current Rate:
Expected Rate:
Total IT Experience [Years]:
Total US Experience [Years]:
Key Skill Set:
Best time to call:
In case you are not interested, I will be very grateful if you can pass this position to your colleagues or friends who might be interested.
All your information will be kept confidential according to EEO guidelines.
$100k-125k yearly est. 22h ago
Staff Data Scientist
Recursion Pharmaceuticals 4.2
Data engineer job in New York, NY
Your work will change lives. Including your own. Please note: Our offices will be closed for our annual winter break from December 22, 2025, to January 2, 2026. Our response to your application will be delayed. The Impact You'll Make As a member of Recursion's AI-driven drug discovery initiatives, you will be at the forefront of reimagining how biological knowledge is generated, stored, accessed, and reasoned upon by LLMs. You will play a key role in developing the biological reasoning infrastructure, connecting large-scale data and codebases with dynamic, agent-driven AI systems.You will be responsible for defining the architecture that grounds our agents in biological truth. This involves integrating biomedical resources to enable AI systems to reason effectively and selecting the most appropriate data retrieval strategies to support those insights. This is a highly collaborative role: you will partner with machine learning engineers, biologists, chemists, and platform teams to build the connective tissue that allows our AI agents to reason like a scientist. The ideal candidate possesses deep expertise in both core bioinformatics/cheminformatics libraries and modern GenAI frameworks (including RAG and MCP), a strong architectural vision, and the ability to translate high-potential prototypes into scalable production workflows.
In this role, you will:
* Architect and maintain robust infrastructure to keep critical internal and external biological resources (e.g., ChEMBL, Ensembl, Reactome, proprietary assays) up-to-date and accessible to reasoning agents.
* Design sophisticated context retrieval strategies, choosing the most effective approach for each biological use case, whether working with structured, entity-focused data, unstructured RAG, or graph-based representations.
* Integrate established bioinformatics/cheminformatics libraries into a GenAI ecosystem, creating interfaces (such as via MCP) that allow agents to autonomously query and manipulate biological data.
* Pilot methods for tool use by LLMs, enabling the system to perform complex tasks like pathway analysis on the fly rather than relying solely on memorized weights.
* Develop scalable, production-grade systems that serve as the backbone for Recursion's automated scientific reasoning capabilities.
* Collaborate cross-functionally with Recursion's core biology, chemistry, data science and engineering teams to ensure our biological data and the reasoning engines are accurately reflecting the complexity of disease biology and drug discovery.
* Present technical trade-offs (e.g., graph vs. vector) to leadership and stakeholders in a clear, compelling way that aligns technical reality with product vision.
The Team You'll Join
You'll join a bold, agile team of scientists and engineers dedicated to building comprehensive biological maps by integrating Recursion's in-house datasets, patient data, and external knowledge layers to enable sophisticated agent-based reasoning. Within this cross-functional team, you will design and maintain the biological context and data structures that allow agents to reason accurately and efficiently. You'll collaborate closely with wet-lab biologists and core platform engineers to develop systems that are not only technically robust but also scientifically rigorous. The ideal candidate is curious about emerging AI technologies, passionate about making biological data both machine-readable and machine-understandable, and brings a strong foundation in systems biology, biomedical data analysis, and agentic AI systems.
The Experience You'll Need
* PhD in a relevant field (Bioinformatics, Cheminformatics, Computational Biology, Computer Science, Systems Biology) with 5+ years of industry experience, or MS in a relevant field with 7+ years of experience, focusing on biological data representation and retrieval.
* Proficiency in utilizing major public biological databases (NCBI, Ensembl, STRING, GO) and using standard bioinformatics/cheminformatics toolkits (e.g., RDKit, samtools, Biopython).
* Strong skills in designing and maintaining automated data pipelines that support continuous ingestion, transformation, and refresh of biological data without manual intervention.
* Ability to work with knowledge graph data models and query languages (e.g., RDF, SPARQL, OWL) and translate graph-structured data into relational or other non-graph representations, with a strong judgment in evaluating trade-offs between different approaches.
* Competence in building and operating GenAI stacks, including RAG systems, vector databases, and optimization of context windows for large-scale LLM deployments.
* Hands-on expertise with agentic AI frameworks (e.g., MCP, Google ADK, LangChain, AutoGPT) and familiarity with leading LLMs (e.g., Google Gemini/Gemma) in agentic workflows, including benchmarking and evaluating agent performance on bioinformatics/cheminformatics tasks such as structure prediction, target identification, and pathway mapping.
* Strong Python skills and adherence to software engineering best practices, including CI/CD, Git-based version control, and modular design.
* Excellent cross-functional communication skills, ability to clearly explain complex architectural decisions to both scientific domain experts and technical stakeholders.
Nice to Have
* Strong background in machine learning and deep learning, including hands-on experience with foundation models and modern neural architectures.
* Fine-tuning LLMs on scientific corpora for domain-specific reasoning.
* Integrating LLMs with experimental or proprietary assay data in live scientific workflows.
* Background in drug discovery and target identification.
* Meaningful contributions to open-source libraries, research codebases, or community-driven tools.
Working Location & Compensation:
This is an office-based, hybrid role in either our Salt Lake City, UT or New York City, NY offices. Employees are expected to work in the office at least 50% of the time.
At Recursion, we believe that every employee should be compensated fairly. Based on the skill and level of experience required for this role, the estimated current annual base range for this role is $200,600 - $238,400. You will also be eligible for an annual bonus and equity compensation, as well as a comprehensive benefits package.
#LI-DNI
The Values We Hope You Share:
* We act boldly with integrity. We are unconstrained in our thinking, take calculated risks, and push boundaries, but never at the expense of ethics, science, or trust.
* We care deeply and engage directly. Caring means holding a deep sense of responsibility and respect - showing up, speaking honestly, and taking action.
* We learn actively and adapt rapidly. Progress comes from doing. We experiment, test, and refine, embracing iteration over perfection.
* We move with urgency because patients are waiting. Speed isn't about rushing but about moving the needle every day.
* We take ownership and accountability. Through ownership and accountability, we enable trust and autonomy-leaders take accountability for decisive action, and teams own outcomes together.
* We are One Recursion. True cross-functional collaboration is about trust, clarity, humility, and impact. Through sharing, we can be greater than the sum of our individual capabilities.
Our values underpin the employee experience at Recursion. They are the character and personality of the company demonstrated through how we communicate, support one another, spend our time, make decisions, and celebrate collectively.
More About Recursion
Recursion (NASDAQ: RXRX) is a clinical stage TechBio company leading the space by decoding biology to radically improve lives. Enabling its mission is the Recursion OS, a platform built across diverse technologies that continuously generate one of the world's largest proprietary biological and chemical datasets. Recursion leverages sophisticated machine-learning algorithms to distill from its dataset a collection of trillions of searchable relationships across biology and chemistry unconstrained by human bias. By commanding massive experimental scale - up to millions of wet lab experiments weekly - and massive computational scale - owning and operating one of the most powerful supercomputers in the world, Recursion is uniting technology, biology and chemistry to advance the future of medicine.
Recursion is headquartered in Salt Lake City, where it is a founding member of BioHive, the Utah life sciences industry collective. Recursion also has offices in Toronto, Montréal, New York, London, Oxford area, and the San Francisco Bay area. Learn more at ****************** or connect on X (formerly Twitter) and LinkedIn.
Recursion is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, veteran status, or any other characteristic protected under applicable federal, state, local, or provincial human rights legislation.
Accommodations are available on request for candidates taking part in all aspects of the selection process.
Recruitment & Staffing Agencies: Recursion Pharmaceuticals and its affiliate companies do not accept resumes from any source other than candidates. The submission of resumes by recruitment or staffing agencies to Recursion or its employees is strictly prohibited unless contacted directly by Recursion's internal Talent Acquisition team. Any resume submitted by an agency in the absence of a signed agreement will automatically become the property of Recursion, and Recursion will not owe any referral or other fees. Our team will communicate directly with candidates who are not represented by an agent or intermediary unless otherwise agreed to prior to interviewing for the job.
$200.6k-238.4k yearly Auto-Apply 12d ago
Lead Data Scientist
Smarsh 4.6
Data engineer job in New York, NY
Who are we? Smarsh empowers its customers to manage risk and unleash intelligence in their digital communications. Our growing community of over 6500 organizations in regulated industries counts on Smarsh every day to help them spot compliance, legal or reputational risks in 80+ communication channels before those risks become regulatory fines or headlines. Relentless innovation has fueled our journey to consistent leadership recognition from analysts like Gartner and Forrester, and our sustained, aggressive growth has landed Smarsh in the annual Inc. 5000 list of fastest-growing American companies since 2008.
Summary
As a Lead Data Scientist (NLP & Financial Compliance) at Smarsh, you will spearhead the development of state-of-the-art natural language processing (NLP) and large language model (LLM) solutions that power next-generation compliance and surveillance systems. You'll work on highly specialized problems at the intersection of natural language processing, communications intelligence, financial supervision, and regulatory compliance, where unstructured data from emails, chats, voice transcripts, and trade communications hold the keys to uncovering misconduct and risk.
The role will involve working with other Senior Data Scientists and mentoring Associate Data Scientists in analyzing complex data, generating insights, and creating solutions as needed across a variety of tools and platforms. This role demands both technical excellence in NLP modeling and a deep understanding of financial domain behavior-including insider trading, market manipulation, off-channel communications, MNPI, bribery, and other supervisory risk areas. The ideal candidate for this position will possess the ability to perform both independent and team-based research and generate insights from large data sets with a hands-on/can do attitude of servicing/managing day to day data requests and analysis.
This role also offers a unique opportunity to get exposure to many problems and solutions associated with taking machine learning and analytics research to production. On any given day, you will have the opportunity to interface with business leaders, machine learning researchers, dataengineers, platform engineers, data scientists and many more, enabling you to level up in true end-to-end data science proficiency.How will you contribute?
Collect, analyze, and interpret small/large datasets to uncover meaningful insights to support the development of statistical methods / machine learning algorithms.
Lead the design, training, and deployment of NLP and transformer-based models for financial surveillance and supervisory use cases (e.g., misconduct detection, market abuse, trade manipulation, insider communication).
Development of machine learning models and other analytics following established workflows, while also looking for optimization and improvement opportunities
Data annotation and quality review
Exploratory data analysis and model fail state analysis
Contribute to model governance, documentation, and explainability frameworks aligned with internal and regulatory AI standards.
Client/prospect guidance in machine learning model and analytic fine-tuning/development processes
Provide guidance to junior team members on model development and EDA
Work with Product Manager(s) to intake project/product requirements and translate these to technical tasks within the team's tooling, technique and procedures
Continued self-led personal development
What will you bring?
Strong understanding of financial markets, compliance, surveillance, supervision, or regulatory technology
Experience with one or more data science and machine/deep learning frameworks and tooling, including scikit-learn, H2O, keras, pytorch, tensorflow, pandas, numpy, carot, tidyverse
Command of data science and statistics principles (regression, Bayes, time series, clustering, P/R, AUROC, exploratory data analysis etc…)
Strong knowledge of key programming concepts (e.g. split-apply-combine, data structures, object-oriented programming)
Solid statistics knowledge (hypothesis testing, ANOVA, chi-square tests, etc…)
Knowledge of NLP transfer learning, including word embedding models (glo Ve, fast Text, word2vec) and transformer models (Bert, SBert, HuggingFace, and GPT-x etc.)
Experience with natural language processing toolkits like NLTK, spa Cy, Nvidia NeMo
Knowledge of microservices architecture and continuous delivery concepts in machine learning and related technologies such as helm, Docker and Kubernetes
Familiarity with Deep Learning techniques for NLP.
Familiarity with LLMs - using ollama & Langchain
Excellent verbal and written skills
Proven collaborator, thriving on teamwork
Preferred Qualifications
Master's or Doctor of Philosophy degree in Computer Science, Applied Math, Statistics, or a scientific field
Familiarity with cloud computing platforms (AWS, GCS, Azure)
Experience with automated supervision/surveillance/compliance tools
$166,000 - $214,000 a year
The above salary range represents Smarsh's good faith and reasonable estimate of the range of possible base compensation at the time of posting.
Any applicable bonus programs will be discussed during the recruiting process.
The salary for this role will be set based on a variety of factors, including but not limited to, internal equity, experience, education, location, specialty and training.
Local cost of living assessments are done for each new hire at the time of offer.
About our culture
Smarsh hires lifelong learners with a passion for innovating with purpose, humility and humor. Collaboration is at the heart of everything we do. We work closely with the most popular communications platforms and the world's leading cloud infrastructure platforms. We use the latest in AI/ML technology to help our customers break new ground at scale. We are a global organization that values diversity, and we believe that providing opportunities for everyone to be their authentic self is key to our success. Smarsh leadership, culture, and commitment to developing our people have all garnered Comparably.com Best Places to Work Awards. Come join us and find out what the best work of your career looks like.
$166k-214k yearly Auto-Apply 12d ago
Data Scientist III (Product Analytics)
Cedar 4.3
Data engineer job in New York, NY
Our healthcare system is the leading cause of personal bankruptcy in the U.S. Every year, over 50 million Americans suffer adverse financial consequences as a result of seeking care, from lower credit scores to garnished wages. The challenge is only getting worse, as high deductible health plans are the fastest growing plan design in the U.S.
Cedar's mission is to leverage data science, smart product design and personalization to make healthcare more affordable and accessible. Today, healthcare providers still engage with its consumers in a “one-size-fits-all” approach; and Cedar is excited to leverage consumer best practices to deliver a superior experience.
Background:
Cedar is a patient payment and engagement platform that leverages data science, AI, consumer-centric design and technology to improve the healthcare experience for patients.
The Data Science team plays a critical role in achieving Cedar's vision. We ensure data and insights are accessible and trusted by all internal and external stakeholders to enable data-driven decision-making. Through analytics, descriptive analyses, and statistical models, we consistently aim to better understand the needs of patients, find improvement opportunities, and impact the product development process. We run experiments to test our various hypotheses and measure our progress. We build ML models and embed them into our web app to deliver a personalized experience to patients. We help differentiate Cedar's product offerings with hard-to-copy competitive advantages.
To do this, we have built an open, supportive, and collaborative team culture. “Love thy teammate” is one of our Data Science values, and we live into this value by operating from a mindset of curiosity, possibility, and the best intentions of each of our team members. As a team, we are dedicated to continually growing our technical excellence and collaborative excellence, equally.
The Role:
Cedar is growing rapidly and the demand for data expertise is increasing exponentially along with our growth. We're seeking a Data Scientist III (Product Analytics) on our Data Science team to work with our product development teams to enable robust data-driven decisions and understand user behavior through in-depth data analysis.
As a Data Scientist III (Product Analytics), you will be responsible for ensuring that we are building world-class products that truly address the needs of our users and customers through data insights. You will identify clear and strategic product opportunities, measure the success of our products and AI personalization platform, understand product strengths and areas for improvement - by immersing yourself in Cedar data and the stories that it tells. You will continuously apply a growth mindset in your approach to problems.
Responsibilities:
Partner with the Product, Design, and Engineering teams to develop an understanding of user behavior and opportunities for product improvement using data
Influence data focused decision making on product teams to move metrics and improve the patient experience
Perform exploratory research to provide product teams with new strategic ideas and levers to improve impactful metrics
Analyze, measure, and track the performance and success of product features
Identify data insights and recommendations for product teams to act and iterate on
Ideate, design, and evaluate product experiments
Build dashboards and visualizations to democratize data insights
Present and clearly communicate data findings and underlying methodology to partners across all levels of the organization
Think from the perspective of a user and develop expert intuition of Cedar's data models and product
Skills and Experience:
5+ years of experience in an analytical role, bonus points for healthcare, fintech, e-commerce, or other related industries
A go-getter attitude, passion for delivering a world-class product experience, and a mindset focused on growth and learning
An analytical attitude and passion for finding opportunities and insights within data
Understanding of statistical methods for research and experimentation purposes
Experience collaborating with product, design, and engineering teams
SQL and Python expertise is a must-have
Experience with dbt, Hex, and Looker is a nice to have
Compensation Range and Benefits
Salary/Hourly Rate Range*: $148,750 - $175,000
This role is equity eligible
This role offers a competitive benefits and wellness package
*Subject to location, experience, and education
#LI- CR1
What do we offer to the ideal candidate?
A chance to improve the U.S. healthcare system at a high-growth company! Our leading healthcare financial platform is scaling rapidly, helping millions of patients per year
Unless stated otherwise, most roles have flexibility to work from home or in the office, depending on what works best for you
For exempt employees: Unlimited PTO for vacation, sick and mental health days-we encourage everyone to take at least 20 days of vacation per year to ensure dedicated time to spend with loved ones, explore, rest and recharge
16 weeks paid parental leave with health benefits for all parents, plus flexible re-entry schedules for returning to work
Diversity initiatives that encourage Cedarians to bring their whole selves to work, including three employee resource groups: be@cedar (for BIPOC-identifying Cedarians and their allies), Pridecones (for LGBTQIA+ Cedarians and their allies) and Cedar Women+ (for female-identifying Cedarians)
Competitive pay, equity (for qualifying roles), and health benefits, including fertility & adoption assistance, that start on the first of the month following your start date (or on your start date if your start date coincides with the first of the month)
Cedar matches 100% of your 401(k) contributions, up to 3% of your annual compensation
Access to hands-on mentorship, employee and management coaching, and a team discretionary budget for learning and development resources to help you grow both professionally and personally
About us
Cedar was co-founded by Florian Otto and Arel Lidow in 2016 after a negative medical billing experience inspired them to help improve our healthcare system. With a commitment to solving billing and patient experience issues, Cedar has become a leading healthcare technology company fueled by remarkable growth. "Over the past several years, we've raised more than $350 million in funding & have the active support of Thrive and Andreessen Horowitz (a16z).
As of November 2024, Cedar is engaging with 26 million patients annually and is on target to process $3.5 billion in patient payments annually. Cedar partners with more than 55 leading healthcare providers and payers including Highmark Inc., Allegheny Health Network, Novant Health, Allina Health and Providence.
$148.8k-175k yearly Auto-Apply 5d ago
Principal Data Scientist : Product to Market (P2M) Optimization
The Gap 4.4
Data engineer job in New York, NY
About Gap Inc. Our brands bridge the gaps we see in the world. Old Navy democratizes style to ensure everyone has access to quality fashion at every price point. Athleta unleashes the potential of every woman, regardless of body size, age or ethnicity. Banana Republic believes in sustainable luxury for all. And Gap inspires the world to bring individuality to modern, responsibly made essentials.
This simple idea-that we all deserve to belong, and on our own terms-is core to who we are as a company and how we make decisions. Our team is made up of thousands of people across the globe who take risks, think big, and do good for our customers, communities, and the planet. Ready to learn fast, create with audacity and lead boldly? Join our team.
About the Role
Gap Inc. is seeking a Principal Data Scientist with deep expertise in operations research and machine learning to lead the design and deployment of advanced analytics solutions across the Product-to-Market (P2M) space. This role focuses on driving enterprise-scale impact through optimization and data science initiatives spanning pricing, inventory, and assortment optimization.
The Principal Data Scientist serves as a senior technical and strategic thought partner, defining solution architectures, influencing product and business decisions, and ensuring that analytical solutions are both technically rigorous and operationally viable. The ideal candidate can lead end-to-end solutioning independently, manage ambiguity and complex stakeholder dynamics, and communicate technical and business risk effectively across teams and leadership levels.
What You'll Do
* Lead the framing, design, and delivery of advanced optimization and machine learning solutions for high-impact retail supply chain challenges.
* Partner with product, engineering, and business leaders to define analytics roadmaps, influence strategic priorities, and align technical investments with business goals.
* Provide technical leadership to other data scientists through mentorship, design reviews, and shared best practices in solution design and production deployment.
* Evaluate and communicate solution risks proactively, grounding recommendations in realistic assessments of data, system readiness, and operational feasibility.
* Evaluate, quantify, and communicate the business impact of deployed solutions using statistical and causal inference methods, ensuring benefit realization is measured rigorously and credibly.
* Serve as a trusted advisor by effectively managing stakeholder expectations, influencing decision-making, and translating analytical outcomes into actionable business insights.
* Drive cross-functional collaboration by working closely with engineering, product management, and business partners to ensure model deployment and adoption success.
* Quantify business benefits from deployed solutions using rigorous statistical and causal inference methods, ensuring that model outcomes translate into measurable value
* Design and implement robust, scalable solutions using Python, SQL, and PySpark on enterprise data platforms such as Databricks and GCP.
* Contribute to the development of enterprise standards for reproducible research, model governance, and analytics quality.
Who You Are
* Master's or Ph.D. in Operations Research, Operations Management, Industrial Engineering, Applied Mathematics, or a closely related quantitative discipline.
* 10+ years of experience developing, deploying, and scaling optimization and data science solutions in retail, supply chain, or similar complex domains.
* Proven track record of delivering production-grade analytical solutions that have influenced business strategy and delivered measurable outcomes.
* Strong expertise in operations research methods, including linear, nonlinear, and mixed-integer programming, stochastic modeling, and simulation.
* Deep technical proficiency in Python, SQL, and PySpark, with experience in optimization and ML libraries such as Pyomo, Gurobi, OR-Tools, scikit-learn, and MLlib.
* Hands-on experience with enterprise platforms such as Databricks and cloud environments
* Demonstrated ability to assess, communicate, and mitigate risk across analytical, technical, and business dimensions.
* Excellent communication and storytelling skills, with a proven ability to convey complex analytical concepts to technical and non-technical audiences.
* Strong collaboration and influence skills, with experience leading cross-functional teams in matrixed organizations.
* Experience managing code quality, CI/CD pipelines, and GitHub-based workflows.
Preferred Qualifications
* Experience shaping and executing multi-year analytics strategies in retail or supply chain domains.
* Proven ability to balance long-term innovation with short-term deliverables.
* Background in agile product development and stakeholder alignment for enterprise-scale initiatives.
Benefits at Gap Inc.
* Merchandise discount for our brands: 50% off regular-priced merchandise at Old Navy, Gap, Banana Republic and Athleta, and 30% off at Outlet for all employees.
* One of the most competitive Paid Time Off plans in the industry.*
* Employees can take up to five "on the clock" hours each month to volunteer at a charity of their choice.*
* Extensive 401(k) plan with company matching for contributions up to four percent of an employee's base pay.*
* Employee stock purchase plan.*
* Medical, dental, vision and life insurance.*
* See more of the benefits we offer.
* For eligible employees
Gap Inc. is an equal-opportunity employer and is committed to providing a workplace free from harassment and discrimination. We are committed to recruiting, hiring, training and promoting qualified people of all backgrounds, and make all employment decisions without regard to any protected status. We have received numerous awards for our long-held commitment to equality and will continue to foster a diverse and inclusive environment of belonging. In 2022, we were recognized by Forbes as one of the World's Best Employers and one of the Best Employers for Diversity.
Salary Range: $201,700 - $267,300 USD
Employee pay will vary based on factors such as qualifications, experience, skill level, competencies and work location. We will meet minimum wage or minimum of the pay range (whichever is higher) based on city, county and state requirements.
$88k-128k yearly est. 37d ago
Data Scientist II - Marketing Mix Models
Walt Disney Co 4.6
Data engineer job in New York, NY
Marketing science - a sub-team within marketing analytics at Disney's Direct to Consumer team (Hulu, Disney+, ESPN+ and Star) - is in search of an econometrician to run marketing mix models (MMM) and associated ancillary analysis. This position will work as part of a team focused primarily on econometric modeling, which also provides support for downstream practices used to inform marketing investment. The analyst plays a hands-on role in modeling efforts. The ideal candidate has a substantial quantitative skill set with direct experience in marketing science practices (MMM, attribution modeling, testing / experimentation, etc.), and should serve as a strong mentor to analysts, helping to onboard new talent in support of wider company goals. Technical acumen as well as narrative-building are integral to the success of this role.
Responsibilities
* Build, sustain and scale econometric models (MMM) for Disney Streaming Services with support from dataengineering and data product teams
* Quantify ROI on marketing investment, determine optimal spend range across the portfolio, identify proposed efficiency caps by channel, set budget amounts and inform subscriber acquisition forecasts
* Support ad hoc strategic analysis to provide recommendations that drive increased return on spend through shifts in mix, flighting, messaging and tactics, and that help cross-validate model results
* Provide insights to marketing and finance teams, helping to design and execute experiments to move recommendations forward based on company goals (e.g., subscriber growth, LTV, etc)
* Support long-term MMM (et.al.) automation, productionalization and scale with support from dataengineering and product
* Build out front-end reporting and dashboarding in partnership with data product analysts and dataengineers to communicate performance metrics across services, markets, channels and subscriber types
Basic Qualifications
* Bachelor's degree in advanced Mathematics, Statistics, Data Science or comparable field of study
* 3+ years of experience in a marketing data science / analytics role with understanding of measurement and optimization best practices
* Coursework or direct experience in applied econometric modeling, ideally in support of measure marketing efficiency and optimize spend, flighting and mix to maximize return on ad spend (i.e., MMM)
* Exposure / understanding of media attribution practices for digital and linear media, the data required to power them and methodologies for measurement
* Understanding of incrementality experiments to validate model recommendations and gain learnings on channel/publisher efficacy
* Exposure to / familiarity with with BI/data concepts and experience building out self-service marketing data solutions
* Strong coding experience in one (or more) data programming languages like Python/R
* Ability to draw insights and conclusions from data to inform model development and business decisions
* Experience in SQL
Preferred Qualifications
* Masters degree in Computer Science, Engineering, Mathematics, Physics, Econometrics, or Statistics
The hiring range for this position in Santa Monica, CA is $117,500 to $157,500 per year and in New York City, NY & Seattle, WA is $123,000 to $165,000. The base pay actually offered will take into account internal equity and also may vary depending on the candidate's geographic region, job-related knowledge, skills, and experience among other factors. A bonus and/or long-term incentive units may be provided as part of the compensation package, in addition to the full range of medical, financial and/or other benefits, dependent on the level and position offered.
About Disney Direct to Consumer:
Disney's Direct to Consumer team oversees the Hulu and Disney+ streaming businesses within Disney Entertainment helping to bring The Walt Disney Company's best-in-class storytelling to fans and families everywhere.
About The Walt Disney Company:
The Walt Disney Company, together with its subsidiaries and affiliates, is a leading diversified international family entertainment and media enterprise that includes three core business segments: Disney Entertainment, ESPN, and Disney Experiences. From humble beginnings as a cartoon studio in the 1920s to its preeminent name in the entertainment industry today, Disney proudly continues its legacy of creating world-class stories and experiences for every member of the family. Disney's stories, characters and experiences reach consumers and guests from every corner of the globe. With operations in more than 40 countries, our employees and cast members work together to create entertainment experiences that are both universally and locally cherished.
This position is with Disney Streaming Services LLC, which is part of a business we call Disney Direct to Consumer.
Disney Streaming Services LLC is an equal opportunity employer. Applicants will receive consideration for employment without regard to race, religion, color, sex, sexual orientation, gender, gender identity, gender expression, national origin, ancestry, age, marital status, military or veteran status, medical condition, genetic information or disability, or any other basis prohibited by federal, state or local law. Disney champions a business environment where ideas and decisions from all people help us grow, innovate, create the best stories and be relevant in a constantly evolving world.
Apply Now Apply Later
Current Employees Apply via My Disney Career
Explore Location
$123k-165k yearly 30d ago
ETL Architect
Integrated Resources 4.5
Data engineer job in New York, NY
A Few Words About Us Integrated Resources, Inc is a premier staffing firm recognized as one of the tri-states most well-respected professional specialty firms. IRI has built its reputation on excellent service and integrity since its inception in 1996. Our mission centers on delivering only the best quality talent, the first time and every time. We provide quality resources in four specialty areas: Information Technology (IT), Clinical Research, Rehabilitation Therapy and Nursing.
Position: ETL Architect
Location: NYC
Duration: 6 months
Job Description:
This opportunity is for individuals who have Hands-on experience in data warehouse design and development. The Role demands more than a typical ETL lead role as it interacts outwardly on projects with architects, PM's, OPS, data modelers, developers, admins, DBA's and testers. This is a hands-on delivery-focused role, and the individual will be responsible for technical delivery of data warehouse and data integration projects
Must have skills
• 7-10 years Hands on experience with Informatica ETL in designing and developing ETL processes based on multiple sources using ETL tools
• Experience in Architecting end to end ETL solutions
• Hands on UNIX experience. Scripting (e.g. shell, perl, alerts, cron, automation)
• Expert at all aspects of relational database design
• Experience working with engineering team with respect to database-related performance tuning, writing of complex SQL, indexing, etc.
Good to Have:
• Experience with IDQ, MDM, other ETL tools
• Experience with dashboard and report development
• Experience with financial services firms will be preferred
Additional Information
Kind Regards
Sachin Gaikwad
Technical Recruiter
Integrated Resources, Inc.
Direct Line : 732-429-1920
$102k-130k yearly est. 60d+ ago
Data Scientist, User Operations
Openai 4.2
Data engineer job in New York, NY
About the Team OpenAI's User Operations organization is building the data and intelligence layer behind AI-assisted operations - the systems that decide when automation should help users, when humans should step in, and how both improve over time. Our flagship platform is transforming customer support into a model for "agent-first" operations across OpenAI.
About the Role
As a Data Scientist on User Operations, you'll design the models, metrics, and experimentation frameworks that power OpenAI's human-AI collaboration loop. You'll build systems that measure quality, optimize automation, and turn operational data into insights that improve product and user experience at scale. You'll partner closely with Support Automation Engineering, Product, and DataEngineering to ensure our data systems are production-grade, trusted, and impactful.
This role is based in San Francisco or New York City. We use a hybrid work model of three days in the office per week and offer relocation assistance to new employees.
Why it matters
Every conversation users have with OpenAI products produces signals about how humans and AI interact. User Ops Data Science turns those signals into insights that shape how we support users today and design agentic systems for tomorrow. This is a unique opportunity to help define how AI collaboration at scale is measured and improved inside OpenAI.
In this role, you will:
* Build and own metrics, classifiers, and data pipelines that determine automation eligibility, effectiveness, and guardrails.
* Design and evaluate experiments that quantify the impact of automation and AI systems on user outcomes like resolution quality and satisfaction.
* Develop predictive and statistical models that improve how OpenAI's support systems automate, measure, and learn from user interactions.
* Partner with engineering and product teams to create feedback loops that continuously improve our AI agents and knowledge systems.
* Translate complex data into clear, actionable insights for leadership and cross-functional stakeholders.
* Develop and socialize dashboards, applications, and other ways of enabling the team and company to answer product data questions in a self-serve way
* Contribute to establishing data science standards and best practices in an AI-native operations environment.
* Partner with other data scientists across the company to share knowledge and continually synthesize learnings across the organization
You might thrive in this role if you have:
* 10+ years of experience in data science roles within product or technology organizations.
* Expertise in statistics and causal inference, applied in both experimentation and observational causal inference studies.
* Expert-level SQL and proficiency in Python for analytics, modeling, and experimentation.
* Proven experience designing and interpreting experiments and making statistically sound recommendations.
* Experience building data systems or pipelines that power production workflows or ML-based decisioning.
* Experience developing and extracting insights from business intelligence tools, such as Mode, Tableau, and Looker.
* Strategic and impact-driven mindset, capable of translating complex business problems into actionable frameworks.
* Ability to build relationships with diverse stakeholders and cultivate strong partnerships.
* Strong communication skills, including the ability to bridge technical and non-technical stakeholders and collaborate across various functions to ensure business impact.
* Ability to operate effectively in a fast-moving, ambiguous environment with limited structure.
* Strong communication skills and the ability to translate complex data into stories for non-technical partners.
Nice-to-haves:
* Familiarity with large language models or AI-assisted operations platforms.
* Experience in operational automation or customer support analytics.
* Background in experimentation infrastructure or human-AI interaction systems.
About OpenAI
OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.
We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic.
For additional information, please see OpenAI's Affirmative Action and Equal Employment Opportunity Policy Statement.
Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer hardware in your possession (including the data contained therein) upon termination of employment or end of assignment; and maintain the confidentiality of proprietary, confidential, and non-public information. In addition, job duties require access to secure and protected information technology systems and related data security obligations.
To notify OpenAI that you believe this job posting is non-compliant, please submit a report through this form. No response will be provided to inquiries unrelated to job posting compliance.
We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.
OpenAI Global Applicant Privacy Policy
At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
$88k-129k yearly est. 26d ago
Sales Data Reporting Consultant
Us Tech Solutions 4.4
Data engineer job in New York, NY
+ Cloud delivers quality computing at scale to provide universal and unfettered access to information, entertainment, and tools. Cloud drives innovation for end-user computing and offers a simple, secure, smart, and constantly improving, self-healing system for a better modern computing experience.
+ The Cloud Go-to-market ("GTM") Strategy & Operations team partners closely with GTM leadership across Sales, Channel, BD, and Marketing to tackle high-priority problems such as market segmentation, use case prioritization, sales resourcing, and sales enablement to drive growth.
**Responsibilities:**
+ Support the Commercial cloud GTM business cadence and generating insights related to business performance to support the team.
+ Lead deep dives: develop, structure and execute analyses required to develop recommendations and communicate findings to senior management.
+ Automate and maintain business insight reports (collecting insights and information from the sales teams).
+ Provide weekly insights into business performance and present to senior management.
+ Maintain and define sales processes, by working with sales and engineering teams.
**Experience:**
+ 4 to 6 years' experience in management consulting, investment banking, corporate strategy, or equivalent project management experience in a technology/media company.
+ Strong business acumen.
+ Ability to pay attention to detail while handling multiple, simultaneous projects.
+ Comfortable working and extracting insights from large sets of data using tools.
+ Sales Force CRM knowledge.
+ SQL knowledge.
+ Strategy and operations experience in tech companies.
US Tech Solutions is a global staff augmentation firm providing a wide range of talent on-demand and total workforce solutions. To know more about US Tech Solutions, please visit *********************** (********************************** .
US Tech Solutions is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, colour, religion, sex, sexual orientation, gender identity, national origin, disability, or status as a protected veteran.
$96k-140k yearly est. 60d+ ago
Data Scientist, Product Analytics
Airtable 4.2
Data engineer job in New York, NY
Airtable is the no-code app platform that empowers people closest to the work to accelerate their most critical business processes. More than 500,000 organizations, including 80% of the Fortune 100, rely on Airtable to transform how work gets done.
Airtable is seeking a product-focused Data Scientist to join our Analytics & Data Science team. In this high-impact role, you'll partner closely with product development teams to transform raw user data into actionable insights that drive growth for Airtable's self-serve business. You'll own critical data pipelines, design and analyze experiments, build dashboards, and deliver strategic insights that inform executive decision-making. This is a unique opportunity to shape the future of a data-driven, AI-native SaaS company and scale analytics best practices across the organization.
What you'll do
Own and maintain core product data pipelines across DBT, Looker, and Omni, ensuring reliability, scalability, and minimal downtime
Build and refine dashboards that deliver self-serve, real-time insights for high-priority product areas
Lead the development and delivery of company-wide strategic insights that connect user behavior patterns and inform executive decisions
Partner with product and engineering teams to define tracking requirements, implement instrumentation, validate data, and deliver launch-specific dashboards or reports
Establish trusted partnerships with product managers, engineers, analysts, and leadership as the go-to resource for product data insights and technical guidance
Collaborate with leadership to define the analytics roadmap, prioritize high-impact initiatives, and assess resource needs for scaling product analytics capabilities
Mentor junior team members and cross-functional partners on analytics best practices and data interpretation; create documentation and training materials to scale institutional knowledge
Support end-to-end analytics for all product launches, including tracking implementation, validation, and post-launch reporting with documented impact measurements
Deliver comprehensive strategic analyses or experiments that connect user behavior patterns and identify new growth opportunities
Lead or participate in cross-functional projects where data science contributions directly influence product or strategy decisions
Migrate engineering team dashboards to Omni or Databricks, enabling self-serve analytics
Who you are
Bachelor's degree in computer science, data science, mathematics/statistics, or related field
6+ years of experience as a data scientist, data analyst, or dataengineer
Experience supporting product development teams and driving product growth insight
Background in SaaS, consumer tech, or data-driven product environments preferred
Expert in SQL and modern data modeling (e.g., dbt, Databricks, Snowflake, BigQuery); sets standards and mentors others on best practices
Deep experience with BI tools and modeling (e.g., Looker, Omni, Hex, Tableau, Mode)
Proficient with experimentation platforms and statistical libraries (e.g., Eppo, Optimizely, LaunchDarkly, scipy, statsmodels)
Proven ability to apply AI/ML tools - from core libraries (scikit-learn, PyTorch, TensorFlow) to GenAI platforms (ChatGPT, Claude, Gemini) and AI-assisted development (Cursor, GitHub Copilot)
Strong statistical foundation; designs and scales experimentation practices that influence product strategy and culture
Translates ambiguous business questions into structured analyses, guiding teams toward actionable insights
Provides thought leadership on user funnels, retention, and growth analytics
Ensures data quality, reliability, and consistency across critical business reporting and analytics workflows
Experience at an AI-native company, with exposure to building or scaling products powered by AI
Knowledge of product analytics tracking frameworks (e.g., Segment, Amplitude, Mixpanel, GA4) and expertise in event taxonomy design
Strong documentation and knowledge-sharing skills; adept at creating technical guides, playbooks, and resources that scale team effectiveness
Models curiosity, creativity, and a learner's mindset; thrives in ambiguity and inspires others to do the same
Crafts compelling narratives with data, aligning stakeholders at all levels and driving clarity in decision-making
Airtable is an equal opportunity employer. We embrace diversity and strive to create a workplace where everyone has an equal opportunity to thrive. We welcome people of different backgrounds, experiences, abilities, and perspectives. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability, protected veteran status or any characteristic protected by applicable federal and state laws, regulations and ordinances. Learn more about your EEO rights as an applicant.
VEVRAA-Federal Contractor
If you have a medical condition, disability, or religious belief/practice which inhibits your ability to participate in any part of the application or interview process, please complete our Accommodations Request Form and let us know how we may assist you. Airtable is committed to participating in the interactive process and providing reasonable accommodations to qualified applicants.
Compensation awarded to successful candidates will vary based on their work location, relevant skills, and experience.
Our total compensation package also includes the opportunity to receive benefits, restricted stock units, and may include incentive compensation. To learn more about our comprehensive benefit offerings, please check out Life at Airtable.
For work locations in the San Francisco Bay Area, Seattle, New York City, and Los Angeles, the base salary range for this role is:$205,200-$266,300 USDFor all other work locations (including remote), the base salary range for this role is:$185,300-$240,000 USD
Please see our Privacy Notice for details regarding Airtable's collection and use of personal information relating to the application and recruitment process by clicking here.
🔒 Stay Safe from Job Scams
All official Airtable communication will come from an @airtable.com email address. We will never ask you to share sensitive information or purchase equipment during the hiring process. If in doubt, contact us at ***************. Learn more about avoiding job scams here.
$205.2k-266.3k yearly Auto-Apply 13d ago
Data Science/Analytics
Sonsoft 3.7
Data engineer job in New York, NY
Sonsoft , Inc. is a USA based corporation duly organized under the laws of the Commonwealth of Georgia. Sonsoft Inc. is growing at a steady pace specializing in the fields of Software Development, Software Consultancy and Information Technology Enabled Services.
Job Description
Domain - Retail and/or CPG domain Knowledge
Retail Execution - Demand Forecasting/Inventory Management/Supply Chain, CRM, Sales and Operation, Procurement
Analytics Knowledge - Clustering, Segmentation, Campaign Analytics, Loyalty Management, Forecasting
Data Understanding of Sales, Inventory, Store, Product and Promotion Data
Ability to translate business requirement to analytical model
Communicate directly with business users
Technical Skills -SQL, Advanced Excel, R/SAS
Nice to have Skill/Competency:
Hadoop, Spark, Python.
Tableau/Spotfire/Qlikview.
Nielsen/IRI/Market Research Data understanding.
Job responsibilities:
Understand business requirement from client and translate it to appropriate Analytics use cases.
Work in a team environment to deliver analytics solution to client.
Define new business metrics as required and build measurement systems as needed.
Enjoy designing, develop and deploy state of the art, data-driven predictive models.
Adept in quick starting optimization options to solve business problems using latest technologies in data mining and statistical modeling,
Advise clients & internal teams on the right analytics algorithm, platforms and approaches to take in addressing complex, open-ended business problems.
Qualifications
Bachelor's degree or foreign equivalent required from an accredited institution. Will also consider three years of progressive experience in the specialty in lieu of every year of education.
At least 7 years of experience with Information Technology or Business Intelligence/Analytics consulting.
Additional Information
** U.S. Citizens and those who are authorized to work independently in the United States are encouraged to apply. We are unable to sponsor at this time.
Note:-
This is a Full-Time & Permanent job opportunity for you.
Only US Citizen, Green Card Holder, GC-EAD, H4-EAD & L2-EAD can apply.
No OPT-EAD, H1B & TN candidates please.
Please mention your Visa Status in your email or resume.
** All your information will be kept confidential according to EEO guidelines.
$68k-96k yearly est. 60d+ ago
ETL Architect/Developer with Informatica or Talend or Abinitio
Cardinal Integrated 4.4
Data engineer job in Monmouth Junction, NJ
Client is currently seeking an ETL Architect/Developer to join the team. They develop, design and deliver high quality business applications, software solutions and services to meet client's business technology requirements. Responsibilities Playing a lead role in software design, architecture, requirements analysis, investigation of leading edge technologies, and software development.
Collaborating with the core development team and the product owner to define and estimate business requirements and then translate these into specific software tasks.
Collaborate as a member of an agile team to get products developed and completed with best in class software development.
Design, build, and maintain high performance, reusable, and reliable ETL packages.
Ensure the best possible performance, quality, and responsiveness of applications developed.
Requirements
Qualifications
5+ years' experience of development of ETL packages.
Expert understanding of the ETL/ELT and architectural principles of data integration and data warehousing
Strong technical knowledge of Enterprise Data Warehouse, database design and Information Management.
Experience with with Informatica or Talend or Abinitio in designing and developing complex ETL packages.
Develop data acquisition, transformation and data load processes
Experience with writing and debugging complex SQL queries
Experience in designing and developing data models
Preferred Qualifications
BSc. or Master's degree in Computer Science or related technical field.
Solid understanding of information management, data modeling, system integration, development methodologies (including unit testing) and web technologies.
Excellent interpersonal and communication skills and an ability to work effectively with teams.
$91k-119k yearly est. 4d ago
Staff Data Scientist, Personalization & Shopping
Pinterest 4.6
Data engineer job in New York, NY
Millions of people around the world come to our platform to find creative ideas, dream about new possibilities and plan for memories that will last a lifetime. At Pinterest, we're on a mission to bring everyone the inspiration to create a life they love, and that starts with the people behind the product.
Discover a career where you ignite innovation for millions, transform passion into growth opportunities, celebrate each other's unique experiences and embrace the flexibility to do your best work. Creating a career you love? It's Possible.
Pinterest is the world's leading visual search and discovery platform, serving over 500 million monthly active users globally on their journey from inspiration to action. At Pinterest, Shopping is a strategic initiative that aims to help Pinners take action by surfacing the most relevant content, at the right time, in the best user-friendly way. We do this through a combination of innovative product interfaces, and sophisticated recommendation systems.
We are looking for a Staff Data Scientist with experience in machine learning and causal inference to help advance Shopping at Pinterest. In your role you will develop methods and models to explain why certain content is being promoted (or not) for a Pinner. You will work in a highly collaborative and cross-functional environment, and be responsible for partnering with Product Managers and Machine Learning Engineers. You are expected to develop a deep understanding of our recommendation system, and generate insights and robust methodologies to answer the "why". The results of your work will influence our development teams, and drive product innovation.
What you'll do:
* Ensure that our recommendation systems produce trustworthy, high-quality outputs to maximize our Pinner's shopping experience.
* Develop robust frameworks, combining online and offline methods, to comprehensively understand the outputs of our recommendations.
* Bring scientific rigor and statistical methods to the challenges of product creation, development and improvement with an appreciation for the behaviors of our Pinners.
* Work cross-functionally to build relationships, proactively communicate key insights, and collaborate closely with product managers, engineers, designers, and researchers to help build the next experiences on Pinterest.
* Relentlessly focus on impact, whether through influencing product strategy, advancing our north star metrics, or improving a critical process.
* Mentor and up-level junior data scientists on the team.
What we're looking for:
* 7+ years of experience analyzing data in a fast-paced, data-driven environment with proven ability to apply scientific methods to solve real-world problems on web-scale data.
* Strong interest and experience in recommendation systems and causal inference.
* Strong quantitative programming (Python/R) and data manipulation skills (SQL/Spark).
* Ability to work independently and drive your own projects.
* Excellent written and communication skills, and able to explain learnings to both technical and non-technical partners.
* A team player eager to partner with cross-functional partners to quickly turn insights into actions.
* Bachelor's/Master's degree in a relevant field such as Computer Science, or equivalent experience.
In-Office Requirement Statement:
* We let the type of work you do guide the collaboration style. That means we're not always working in an office, but we continue to gather for key moments of collaboration and connection.
* This role will need to be in the office for in-person collaboration 1-2 times/quarter and therefore can be situated anywhere in the country.
Relocation Statement:
* This position is not eligible for relocation assistance. Visit our PinFlex page to learn more about our working model.
#LI-REMOTE
#LI-NM4
At Pinterest we believe the workplace should be equitable, inclusive, and inspiring for every employee. In an effort to provide greater transparency, we are sharing the base salary range for this position. The position is also eligible for equity. Final salary is based on a number of factors including location, travel, relevant prior experience, or particular skills and expertise.
Information regarding the culture at Pinterest and benefits available for this position can be found here.
US based applicants only
$164,695-$339,078 USD
Our Commitment to Inclusion:
Pinterest is an equal opportunity employer and makes employment decisions on the basis of merit. We want to have the best qualified people in every job. All qualified applicants will receive consideration for employment without regard to race, color, ancestry, national origin, religion or religious creed, sex (including pregnancy, childbirth, or related medical conditions), sexual orientation, gender, gender identity, gender expression, age, marital status, status as a protected veteran, physical or mental disability, medical condition, genetic information or characteristics (or those of a family member) or any other consideration made unlawful by applicable federal, state or local laws. We also consider qualified applicants regardless of criminal histories, consistent with legal requirements. If you require a medical or religious accommodation during the job application process, please complete this form for support.
$104k-140k yearly est. Auto-Apply 60d+ ago
SAP BRIM Data Migration Consultant
Data-Core System, Inc. 4.2
Data engineer job in Middletown, PA
Job DescriptionSalary: DOE
Data-Core Systems, Inc. is a provider of information technology, consulting, and business process services. We offer breakthrough tech solutions and have worked with companies, hospitals, universities, and government organizations. A proven partner with a passion for client satisfaction, we combine technology innovation, business process expertise, and a global, collaborative workforce that exemplifies the future of work. For more information about Data-Core Systems, Inc., please visit *****************************
Our client is a roadway system, and as a part of their digital transformation, they are implementing a solution based on SAP BRIM & Microsoft Dynamics CE.
Data-Core Systems Inc. is seeking an SAP BRIM Data Migration Consultant to join our Consulting team. You will participate and effectively contribute to the design, development, and implementation of complex applications, often using new technologies. You will provide technical expertise and systems design for individual initiatives. You will have the opportunity to work with other SME consultants from our existing team.
Roles & Responsibilities:
Migrate data from external sources to SAP BRIM
Support Full Life Cycle Migration, including mapping to new systems
Ensure accuracy and integrity of data and applications through analysis, coding, documenting, testing and problem solving
Set up the development environment and configuration of the development tools
Contribute to Cloud Operations for running application systems on a large scale
Required Skills & Experience:
Minimum 8-12 years of relevant experience is required
5+ years of hands-on experience in SAP BRIM
Knowledge in Data Profiling and Data Validation
Bachelors degree in engineering or a bachelors degree in technology from a recognized university
We are an equal opportunity employer.
$78k-106k yearly est. 14d ago
ETL Talend MDM Architect
TRG 4.6
Data engineer job in New York, NY
Responsibilities:
• Develop and test Extract, Transformation, and Loading (ETL) modules based on design specifications
• Develop and test ETL Mappings in Talend
• Plan, test, and deploy ETL mappings, and database code as part of application build process across the enterprise
• Provide effective communications with all levels of internal and external customers and staff
• Must demonstrate knowledge in the following areas:
o Data Integration
o Data Architecture
o Team Lead experience is a plus
• Understand, analyze, assess and recommend ETL environment from technology strategy and operational standpoint
• Understand and assess source system data issues and recommend solution from data integration standpoint
• Create high level, low level technical design documents for data integration
• Design exceptions handling, audit and data resolution processes
• Performance tune ETL environment
• Conduct proof of concepts
• Estimation of work based on functional requirements documents
• Identify system deficiencies and recommending solutions
• Designing, coding, and writing unit test cases from functional requirements
• Delivering efficient and bug-free ETL packages and documentation
• Maintenance and support of enterprise ETL jobs
• Experience with Talend Hadoop tools is a plus
Basic Qualifications:
• 3+ years of development experience on Talend ETL tools
• 7+ years working with one or more of the following ETL Tools: Talend, Informatica, Ab Initio or Data Stage
• 7+ years proficient experience as a developer
• Bachelor's Degree in Computer Science or equivalent
• Database (Oracle, SQL Server, DB2)
• Database Programming (Complex SQL, PL/SQL development knowledge)
• Data Modeling
• Business Analysis
• Top level performer with ability to work independently in short time frames
• Proficient working in a Linux environment
• Experience in scripting languages (Shell, Python or Perl)
• 5+ years of experience deploying large scale projects ETL projects that
• 3+ years of experience in a development lead position
• Data analysis, data mapping, data loading, and data validation
• Understand reusability, parameterization, workflow design, etc.
• Thorough understanding of Entire life cycle of Software and various Software Engineering Methodologies
• Performance tuning of interfaces that extract, transform and load tens of millions of records
• Knowledge of Hadoop ecosystem technologies is a plus
Additional Information
If you are comfortable with the position and location then please revert me back at the earliest with your updated resume and following details or I would really appreciate if you can call me back on my number.
Full Name:
Email:
Skype id:
Contact Nos.:
Current Location:
Open to relocate:
Start Availability:
Work Permit:
Flexible time for INTERVIEW:
Current Company:
Current Rate:
Expected Rate:
Total IT Experience [Years]:
Total US Experience [Years]:
Key Skill Set:
Best time to call:
In case you are not interested, I will be very grateful if you can pass this position to your colleagues or friends who might be interested.
All your information will be kept confidential according to EEO guidelines.
$100k-125k yearly est. 60d+ ago
Marketing Data Science Consultant
Us Tech Solutions 4.4
Data engineer job in New York, NY
+ The Marketing Data Science Associate conducts analysis related to incremental impact of marketing efforts and develops methods of optimization for the Marketing team. + The candidate must have a passion for this type of marketing-specific data science.
+ This role will be responsible for working with internal and external data sets, tools and vendors, partnering with the campaign analyst to develop methodologies and frameworks, and build data science models and data pipelines to ensure accuracy when answering business questions around the marketing business.
**Responsibilities:**
+ Manage data pipelining, development of data tables, creation of relevant script/code and implementation of dashboards.
+ Analyze and develop explanatory models and predictive models using appropriate mathematical methods, helping to inform outcomes-based marketing plans.
+ Execute models and apply statistical tools to enable marketing optimization/selection and measurements - i.e. criteria selection scorecards.
+ Build relationships with cross-functional partners.
+ Support with framing and translating business questions into appropriate action plans for the team.
**Experience:**
+ Experience with client-side media and marketing data science.
+ Experience developing media measurement methodologies, such as marketing mix models, multi-touch attribution logic, and causal lift studies.
+ Understanding of incrementality measurement methodologies and frameworks.
+ Demonstrated skills in project management, stakeholder management, and process flow (Desired).
+ Proactive and flexible, able to operate in a fast-paced, constantly evolving and collaborative environment (Desired).
**Skills:**
+ Marketing Data Science.
+ Incrementality measurement methodologies.
+ Data Analytics.
**Education:**
+ BA/BS degree or equivalent practical experience 2 years of data science and/or data analytics experience.
**About US Tech Solutions:**
US Tech Solutions is a global staff augmentation firm providing a wide range of talent on-demand and total workforce solutions. To know more about US Tech Solutions, please visit *********************** (********************************** .
US Tech Solutions is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, colour, religion, sex, sexual orientation, gender identity, national origin, disability, or status as a protected veteran.
$96k-140k yearly est. 60d+ ago
Microsoft Dynamics 365 CE Data Migration Consultant
Data-Core System, Inc. 4.2
Data engineer job in Middletown, PA
Job DescriptionSalary:
Data-Core Systems, Inc. is a provider of information technology, consulting and business process services. We offer breakthrough tech solutions and have worked with companies, hospitals, universities and government organizations. A proven partner with a passion for client satisfaction, we combine technology innovation, business process expertise and a global, collaborative workforce that exemplifies the future of work. For more information about Data-Core Systems, Inc., please visit*****************************
Our client is a roadway system and as a part of their digital transformation they are implementing a solution based on SAP BRIM & Microsoft Dynamics CE.
Data-Core Systems Inc. is seeking Microsoft Dynamics 365 CE Data Migration Consultantto be a part of our Consulting team. You will be responsible for planning, designing, and executing the migration of customer, account, vehicle, financial, and transaction data from a variety of source systemsincluding legacy CRMs, ERPs, SQL databases, flat files, Excel, cloud platforms, and tolling systemsinto Microsoft Dynamics 365 Customer Engagement (CE). This role involves understanding complex data models, extracting structured and unstructured data, transforming and mapping it to Dynamics CE entities, and ensuring data quality, integrity, and reconciliation throughout the migration lifecycle.
Roles & Responsibilities:
Analyze source system data structures, including customer profiles, accounts, vehicles, transponders, payment methods, transactions, violations, invoices, and billing records
Identify critical data relationships, parent/child hierarchies, and foreign key dependencies
Develop detailed data mapping and transformation documentation from source systems to Dynamics 365 CE entities (standard and custom)
Build, test, and execute ETL pipelines using tools such as SSIS/KingswaySoft, Azure Data Factory, Power Platform Dataflows, or custom .NET utilities
Perform data cleansing, normalization, deduplication, and standardization to meet Dynamics CE data model requirements
Execute multiple migration cycles, including test loads, validation, and final production migration
Ensure referential integrity, high data quality, and accuracy of historical data
Generate reconciliation reports, resolve data inconsistencies, and troubleshoot migration errors
Document migration strategies, execution runbooks, and transformation rules for future reference
Required Skills & Experience:
8-12 years of proven experience migrating data from tolling systems, transportation platforms, legacy CRMs, or other high-volume transactional systems
Strong SQL skills for complex queries, stored procedures, data transformation, and data validation
Hands-on experience with Microsoft Dynamics 365 CE / CRM data model, entities, and relationships
Proficiency with ETL/migration tools: SSIS with KingswaySoft, Azure Data Factory, Power Platform Dataflows, Custom C#/.NET migration scripts
Experience with large-scale migrations involving millions of records
Strong understanding of relational data structures such as: Customer Account Vehicle Transponder Transaction
Ability to analyze large datasets, identify anomalies, and resolve inconsistencies
Bachelors degree in engineering or a bachelors degree in technology from a recognized university
Preferred Skills & Experience:
Experience with financial transactions, billing data, or violation/enforcement records.
Experience in enterprise-scale Dynamics 365 CE migrations.
Familiarity with data governance, security, and compliance requirements for financial or transportation data.
Knowledge of historical data migration and archival strategies.
We are an equal opportunity employer.
The average data engineer in Brick, NJ earns between $71,000 and $128,000 annually. This compares to the national average data engineer range of $80,000 to $149,000.