Data Scientist - Analytics roles draw analytical talent hunting for roles.
Data engineer job in Commack, NY
About the Role We believe that the best decisions are backed by data. We are seeking a curious and analytical Data Scientist to champion our data -driven culture.
In this role, you will act as a bridge between technical data and business strategy. You will mine massive datasets, build predictive models, and-most importantly-tell the story behind the numbers to help our leadership team make smarter choices. You are perfect for this role if you are as comfortable with SQL queries as you are with slide decks.
### What You Will Do
Exploratory Analysis: Dive deep into raw data to discover trends, patterns, and anomalies that others miss.
Predictive Modeling: Build and test statistical models (Regression, Time -series, Clustering) to forecast business outcomes and customer behavior.
Data Visualization: Create clear, impactful dashboards using Tableau, PowerBI, or Python libraries (Matplotlib/Seaborn) to visualize success metrics.
Experimentation: Design and analyze A/B tests to optimize product features and marketing campaigns.
Data Cleaning: Work with Data Engineers to clean and structure messy data for analysis.
Strategy: Present findings to stakeholders, translating complex math into clear, actionable business recommendations.
Requirements
Experience: 2+ years of experience in Data Science or Advanced Analytics.
The Toolkit: Expert proficiency in Python or R for statistical analysis.
Data Querying: Advanced SQL skills are non -negotiable (Joins, Window Functions, CTEs).
Math Mindset: Strong grasp of statistics (Hypothesis testing, distributions, probability).
Visualization: Ability to communicate data visually using Tableau, PowerBI, or Looker.
Communication: Excellent verbal and written skills; you can explain a p -value to a non -technical manager.
### Preferred Tech Stack (Keywords)
Languages: Python (Pandas, NumPy), R, SQL
Viz Tools: Tableau, PowerBI, Looker, Plotly
Machine Learning: Scikit -learn, XGBoost (applied to business problems)
Big Data: Spark, Hadoop, Snowflake
Benefits
Salary Range: $50,000 - $180,000 USD / year (Commensurate with location and experience)
Remote Friendly: Work from where you are most productive.
Learning Budget: Stipend for data courses (Coursera, DataCamp) and books.
Senior Data Engineer
Data engineer job in Farmingdale, NY
D'Addario & Company is the world's largest manufacturer and distributor of musical instrument accessories. As a U.S.-based manufacturing leader, we pride ourselves on high-automation machinery, cutting-edge technology, and a deep commitment to environmentally sustainable practices. Most importantly, we're proud of our diverse team of individuals who embody our core values-family, curiosity, passion, candor, and responsibility-and bring them to life every day.
D'Addario is seeking a Senior Data Engineer to help architect, build, and optimize the next generation of our global data infrastructure. In this role, you'll design and maintain production-grade data pipelines, support AI and machine learning initiatives, and serve as a technical mentor within a growing Business Intelligence team. You'll work closely with the Global Director of BI to deliver scalable solutions that power insights and innovation across the organization. This position is ideal for someone who thrives on solving complex data challenges, enjoys bringing structure to large datasets, and is passionate about enabling smarter decision-making through data.
This is a hybrid role and will require the candidate to work on-site in the Farmingdale office three days a week.
At D'Addario, we don't just offer a job-we offer a career with one of the most iconic names in the music industry. We're passionate about innovation, craftsmanship, and creating a workplace where diverse backgrounds, perspectives, and ideas thrive. We're eager to connect with individuals who bring fresh thinking and a collaborative spirit. If you're ready to make an impact, we'd love to hear how you'll add value to our team.
Some Perks & Benefits of Working at D'Addario:
Competitive compensation package
Health, vision, and dental insurance
12 weeks of fully paid parental leave
Fertility and family-building benefits
401(k) retirement plan with generous employer contributions
Career pathing and professional development via LinkedIn Learning
Paid Time Off (PTO) and flexible sick day policy
12 Paid Holidays
Life and AD&D Insurance
Enhanced Short-Term Disability Insurance
Employee Assistance Program (EAP)
Tuition Reimbursement
Discounts on D'Addario products and merchandise
Company jam nights, artist performances, holiday parties, and special events
A passionate, talented team that loves what they do!
Responsibilities
Build & Optimize Pipelines: Design, implement, and maintain robust, high-performance data pipelines to support analytical models within Microsoft Fabric and our data environment.
Data Integration: Connect and harmonize new data sources, including ERP, e-commerce platforms, and external APIs.
Mentorship & Standards: Guide junior BI team members, lead code reviews, and establish coding, documentation, and testing best practices.
AI/ML Enablement: Partner on machine learning and AI projects from proof-of-concept through deployment, supporting predictive and prescriptive analytics into production workflows.
Advanced Analytics Development: Team up with analysts to prepare data products and predictive models using Python, PySpark, and modern ML frameworks.
Collaboration: Work with stakeholders across Sales, Marketing, Operations, and Product to translate business requirements and align priorities.
Technical Leadership: Drive data engineering excellence through continuous improvement, quality assurance, and innovation in data architecture and governance.
Qualifications
5+ years of experience building and maintaining production-grade data pipelines.
Advanced programming skills in Python, PySpark, and SQL.
Strong background in data modeling and scalable analytics.
Experience deploying machine learning models and data products in production environments.
Solid understanding of cloud data platforms (Azure preferred).
Bachelor's degree in Computer Science, Engineering, Data Science, or equivalent experience.
Clear communicator with the ability to simplify complex technical concepts.
Proven leadership in mentoring and developing technical talent.
Highly organized, self-directed, and comfortable in fast-paced environments.
Passion for using data to drive innovation and business impact.
The base salary range for this role would be commensurate with experience: $140k to $165k per year
#LI-HYBRID
Auto-ApplyHadoop Developer
Data engineer job in Lake Success, NY
A big opportunity! Great pay!
Love Hadoop? Do you have it in you to become a mentor? Then, my client requires YOUR mojo with them.
This is an excellent opportunity for those aspiring for a growth oriented company. You will be the first person to work on Hadoop in their New York office. You will be responsible for starting and educating new team members on Hadoop technologies, with a combination of hands on and training responsibilities.
Interesting?? Scroll down for more..
Job Description
Not an exhaustive set of responsibilities, but an overview:
You will be responsible for Architect, design and develop code that consistently adheres to functional programming principles.
You will design, develop, and maintain high volume data processing batch/streaming jobs using industry standard tools and frameworks in Hadoop ecosystem.
You will help my client continue as an industry leader by exploring new technologies,languages and techniques in the rapidly evolving world of high volume data processing.
You will collaborate with team members using Agile techniques, including: pair programming, test driven development (TDD), code reviews, and retrospectives.
If this has piqued your curiosity, please check if YOU have the following:
Experience in Hadoop Ecosystems (MapReduce/Spark/Hive/Oozie/scoop etc...)
Development experience in one of the following - Java, Python OR Scala.
Capable of working with and influencing large, diverse teams.
Experience working with large data sets and high volume data processing.
Here's what we can offer:
A competitive Base Salary of $140K with a Bonus of up to $10K and Benefits.
Note: We are looking for local candidates.
Additional Information
All your information will be kept confidential according to EEO guidelines.
Ping me at ******************** to know more.
Easy ApplyJunior Data Scientist
Data engineer job in New Haven, CT
About Us
Bexorg is revolutionizing drug discovery by restoring molecular activity in postmortem human brains. Through our BrainEx platform, we directly experiment on functionally preserved human brain tissue, creating enormous high-fidelity molecular datasets that fuel AI-driven breakthroughs in treating CNS diseases. We are looking for a Junior Data Scientist to join our team and dive into this one-of-a-kind data. In this onsite role, you will work at the intersection of computational biology and machine learning, helping analyze high-dimensional brain data and uncover patterns that could lead to the next generation of CNS therapeutics. This is an ideal opportunity for a recent graduate or early-career scientist to grow in a fast-paced, mission-driven environment.
The Job
Data Analysis & Exploration: Work with large-scale molecular datasets from our BrainEx experiments - including transcriptomic, proteomic, and metabolic data. Clean, transform, and explore these high-dimensional datasets to understand their structure and identify initial insights or anomalies.
Collaborative Research Support: Collaborate closely with our life sciences, computational biology and deep learning teams to support ongoing research. You will help biologists interpret data results and assist machine learning researchers in preparing data for modeling, ensuring that domain knowledge and data science intersect effectively.
Machine Learning Model Execution: Run and tune machine learning and deep learning models on real-world central nervous system (CNS) data. You'll help set up experiments, execute training routines (for example, using scikit-learn or PyTorch models), and evaluate model performance to extract meaningful patterns that could inform drug discovery.
Statistical Insight Generation: Apply statistical analysis and visualization techniques to derive actionable insights from complex data. Whether it's identifying gene expression patterns or correlating molecular changes with experimental conditions, you will contribute to turning data into scientific discoveries.
Reporting & Communication: Document your analysis workflows and results in clear reports or dashboards. Present findings to the team, highlighting key insights and recommendations. You will play a key role in translating data into stories that drive decision-making in our R&D efforts.
Qualifications and Skills:
Strong Python Proficiency: Expert coding skills in Python and deep familiarity with the standard data science stack. You have hands-on experience with NumPy, pandas, and Matplotlib for data manipulation and visualization; scikit-learn for machine learning; and preferably PyTorch (or similar frameworks like TensorFlow) for deep learning tasks.
Educational Background: A Bachelor's or Master's degree in Data Science, Computer Science, Computational Biology, Bioinformatics, Statistics, or a related field. Equivalent practical project experience or internships in data science will also be considered.
Machine Learning Knowledge: Solid understanding of machine learning fundamentals and algorithms. Experience developing or applying models to real or simulated datasets (through coursework or projects) is expected. Familiarity with high-dimensional data techniques or bioinformatics methods is a plus.
Analytical & Problem-Solving Skills: Comfortable with statistics and data analysis techniques for finding signals in noisy data. Able to break down complex problems, experiment with solutions, and clearly interpret the results.
Team Player: Excellent communication and collaboration skills. Willingness to learn from senior scientists and ability to contribute effectively in a multidisciplinary team that includes biologists, data engineers, and AI researchers.
Motivation and Curiosity: Highly motivated, with an evident passion for data-driven discovery. You are excited by Bexorg's mission and eager to take on challenging tasks - whether it's mastering a new analysis method or digging into scientific literature - to push our research forward.
Local to New Haven, CT preferred. No relocation offered for this position.
Bexorg is an equal opportunity employer. We strive to create a supportive and inclusive workplace where contributions are valued and celebrated, and our employees thrive by being themselves and are inspired to do their best work. We seek applicants of all backgrounds and identities, across race, color, ethnicity, national origin or ancestry, citizenship, religion, sex, sexual orientation, gender identity or expression, veteran status, marital status, pregnancy or parental status, or disability. Applicants will not be discriminated against based on these or other protected categories or social identities. Bexorg will also consider for employment qualified applicants with criminal histories in a manner consistent with applicable federal, state and local law.
Data Engineer
Data engineer job in New Hyde Park, NY
Job Description
Data is pivotal to our goal of frequent launch and rapid iteration. We're recruiting a Data Engineer at iRocket to build pipelines, analytics, and tools that support propulsion test, launch operations, manufacturing, and vehicle performance.
The Role
Design and build data pipelines for test stands, manufacturing machines, launch telemetry, and operations systems.
Develop dashboards, real-time monitoring, data-driven anomaly detection, performance trending, and predictive maintenance tools.
Work with engineers across propulsion, manufacturing, and operations to translate data-needs into data-products.
Maintain data architecture, ETL processes, cloud/edge-data systems, and analytics tooling.
Support A/B testing, performance metrics, and feed insights back into design/manufacturing cycles.
Requirements
Bachelor's degree in Computer Science, Data Engineering, or related technical field.
2+ years of experience building data pipelines, ETL/ELT workflows, and analytics systems.
Proficient in Python, SQL, cloud data platforms (AWS, GCP, Azure), streaming/real-time analytics, and dashboarding (e.g., Tableau, PowerBI).
Strong ability to work cross-functionally and deliver data-products to engineering and operations teams.
Strong communication, documentation, and a curiosity-driven mindset.
Benefits
Health Care Plan (Medical, Dental & Vision)
Retirement Plan (401k, IRA)
Life Insurance (Basic, Voluntary & AD&D)
Paid Time Off (Vacation, Sick & Public Holidays)
Family Leave (Maternity, Paternity)
Short Term & Long Term Disability
Wellness Resources
Hadoop Developer - II
Data engineer job in Lake Success, NY
The sheer scale of our capabilities and client engagements and the way we collaborate, operate and deliver value provides an unparalleled opportunity to grow and advance. Choose Workila, and make delivering innovative work part of your extraordinary career.
Job Description
MUST-HAVES
Development exp in Hadoop Ecosystems (MapReduce/Spark/Hive/Oozie/Sqoop etc)
Development experience in one of the following - Java, Python OR Scala
Capable of working with and influencing large diverse teams. Provide example
7+ yr overall development experience (2+ in Hadoop min.)
Effective communication and proven leadership/mentoring experience
NICE-TO-HAVES
Have worked with large datasets and high volume data processing
ROLE DESCRIPTION
DESCRIPTION OF DUTIES
The Hadoop Developer will be responsible to Architect, design and develop code that consistently adheres to functional programming principles. The successful candidate will also have a balance of hands on work with training and mentoring of a team. They will also be an effective communicator. Other responsibilities include:
Design, develop, and maintain high volume data processing batch/streaming jobs using industry standard tools and frameworks in Hadoop ecosystem.
Maintain our position as industry leader by exploring new technologies,languages and techniques in the rapidly evolving world of high volume data processing.
Collaborate with team members using Agile techniques, including: pair programming, test driven development (TDD), code reviews, and retrospectives.
CULTURE AND PERKS
The Hadoop Developer will be the first person to work on Hadoop in our Lake Success Office. They will be responsible for starting and educating new team members on Hadoop technologies, with a combination of hands on and training responsibilities.
Our company is a dynamic, innovative technology company that revolutionized the automotive retail industry with the first online finance and credit application network in 2001.
Our state-of-the-art, web-based solutions are embraced by all major segments of the automotive retailing trade including dealers, financing sources, original equipment manufacturers (OEMs), third-party retailers, agents and aftermarket providers, fueling our tremendous growth.
We have location-specific programs and amenities that keep team members energized, engaged and doing great work, including:
Health club partial reimbursement
Options for discounts on car purchases with our vendor partners
Local business partnership programs
Volunteer Paid time
Massages
Company Picnic
Company Party
Hackathons
Holiday-inspired Company days
Shuttle Service to and from LIRR station to our building in Lake Success, NY
MORE INFORMATION
Full-time Mid-Level Computer Software
SALARY RANGE $110,000 - $130,000
PERFORMANCE BONUS $0 - $10,000, Bonus Plan
SIGNING BONUS None
BENEFITS
Medical Ins.
Dental Ins.
Other
OPEN UNTIL (MAY 24)
RELOCATION Not offered
DIRECT REPORTS Zero
REPORTS TO Director of Business Intelligence
REMOTE WORK Remote work not available
TRAVEL Travel not required
VISA Candidate visas are supported
Additional Information
Apply online by clicking on green label "I am Interested"
or call if you have any question however applying online is the best way to apply.
US # ************
India # 9999 883 470
Tech Lead, Data & Inference Engineer
Data engineer job in Stamford, CT
Our Client
A fast moving and venture backed advertising technology startup based in San Francisco. They have raised twelve million dollars in funding and are transforming how business to business marketers reach their ideal customers. Their identity resolution technology blends business and consumer signals to convert static audience lists into high match and cross channel segments without the use of cookies. By transforming first party and third party data into precision targetable audiences across platforms such as Meta, Google and YouTube, they enable marketing teams to reach higher match rates, reduce wasted advertising spend and accelerate pipeline growth. With a strong understanding of how business buyers behave in channels that have traditionally been focused on business to consumer activity, they are redefining how business brands scale demand generation and account based efforts.
About Us
Catalyst Labs is a leading talent agency with a specialized vertical in Applied AI, Machine Learning, and Data Science. We stand out as an agency thats deeply embedded in our clients recruitment operations.
We collaborate directly with Founders, CTOs, and Heads of AI in those themes who are driving the next wave of applied intelligence from model optimization to productized AI workflows. We take pride in facilitating conversations that align with your technical expertise, creative problem-solving mindset, and long-term growth trajectory in the evolving world of intelligent systems.
Location: San Francisco
Work type: Full Time,
Compensation: above market base + bonus + equity
Roles & Responsibilities
Lead the design, development and scaling of an end to end data platform from ingestion to insights, ensuring that data is fast, reliable and ready for business use.
Build and maintain scalable batch and streaming pipelines, transforming diverse data sources and third party application programming interfaces into trusted and low latency systems.
Take full ownership of reliability, cost and service level objectives. This includes achieving ninety nine point nine percent uptime, maintaining minutes level latency and optimizing cost per terabyte. Conduct root cause analysis and provide long lasting solutions.
Operate inference pipelines that enhance and enrich data. This includes enrichment, scoring and quality assurance using large language models and retrieval augmented generation. Manage version control, caching and evaluation loops.
Work across teams to deliver data as a product through the creation of clear data contracts, ownership models, lifecycle processes and usage based decision making.
Guide architectural decisions across the data lake and the entire pipeline stack. Document lineage, trade offs and reversibility while making practical decisions on whether to build internally or buy externally.
Scale integration with application programming interfaces and internal services while ensuring data consistency, high data quality and support for both real time and batch oriented use cases.
Mentor engineers, review code and raise the overall technical standard across teams. Promote data driven best practices throughout the organization.
Qualifications
Bachelors or Masters degree in Computer Science, Computer Engineering, Electrical Engineering, or Mathematics.
Excellent written and verbal communication; proactive and collaborative mindset.
Comfortable in hybrid or distributed environments with strong ownership and accountability.
A founder-level bias for actionable to identify bottlenecks, automate workflows, and iterate rapidly based on measurable outcomes.
Demonstrated ability to teach, mentor, and document technical decisions and schemas clearly.
Core Experience
6 to 12 years of experience building and scaling production-grade data systems, with deep expertise in data architecture, modeling, and pipeline design.
Expert SQL (query optimization on large datasets) and Python skills.
Hands-on experience with distributed data technologies (Spark, Flink, Kafka) and modern orchestration tools (Airflow, Dagster, Prefect).
Familiarity with dbt, DuckDB, and the modern data stack; experience with IaC, CI/CD, and observability.
Exposure to Kubernetes and cloud infrastructure (AWS, GCP, or Azure).
Bonus: Strong Node.js skills for faster onboarding and system integration.
Previous experience at a high-growth startup (10 to 200 people) or early-stage environment with a strong product mindset.
Senior Market Data Engineer
Data engineer job in Old Greenwich, CT
WorldQuant develops and deploys systematic financial strategies across a broad range of asset classes and global markets. We seek to produce high-quality predictive signals (alphas) through our proprietary research platform to employ financial strategies focused on market inefficiencies. Our teams work collaboratively to drive the production of alphas and financial strategies - the foundation of a balanced, global investment platform.
WorldQuant is built on a culture that pairs academic sensibility with accountability for results. Employees are encouraged to think openly about problems, balancing intellectualism and practicality. Excellent ideas come from anyone, anywhere. Employees are encouraged to challenge conventional thinking and possess an attitude of continuous improvement.
Our goal is to hire the best and the brightest. We value intellectual horsepower first and foremost, and people who demonstrate an outstanding talent. There is no roadmap to future success, so we need people who can help us build it.
Technologists at WorldQuant research, design, code, test and deploy firmwide platforms and tooling while working collaboratively with researchers and portfolio managers. Our environment is relaxed yet intellectually driven. We seek people who think in code and are motivated by being around like-minded people.
The Role:
* Design and build real-time market data processing systems covering global markets and multiple asset classes
* Architect and implement high-performance software solutions for processing market data feeds at scale
* Drive technical innovation by leveraging emerging technologies to enhance system telemetry, monitoring, and operational efficiency
* Provide technical leadership and escalation support for production market data systems
* Analyze system performance and design data-driven approaches to optimize market data processing workflows
* Lead the design of data governance systems for tracking availability, access patterns, and usage metrics
What You Will Bring:
* Degree in a quantitative or technical discipline from top university and strong academic scores
* Expert-level C++ proficiency with demonstrated experience in other object-oriented languages (Java, C#)
* Experience with scripting languages such as Perl, Python, and shell scripting for automation and data processing
* Deep experience with tick-by-tick market data processing, including data normalization, feed handling, and real-time analytic
* Excellent communication skills with ability to collaborate effectively across technical and business teams
* Have experience working under Linux environment
Our Benefits:
* Core Benefits: Fully paid medical and dental insurance for employees and dependents, flexible spending account, 401k, fully paid parental leave, generous PTO (paid time off) that consists of:
* twenty vacation days that are pro-rated based on the employee's start date, at an accrual of 1.67 days per month,
* three personal days, and
* ten sick days.
* Perks: Employee discounts for gym memberships, wellness activities, healthy snacks, casual dress code
* Training: learning and development courses, speakers, team-building off-site
* Employee resource groups
Pay Transparency:
WorldQuant is a total compensation organization where you will be eligible for a base salary, discretionary performance bonus, and benefits.
To provide greater transparency to candidates, we share base pay ranges for all US-based job postings regardless of state. We set standard base pay ranges for all roles based on job function and level, benchmarked against similar stage organizations. When finalizing an offer, we will take into consideration an individual's experience level and the qualifications they bring to the role to formulate a competitive total compensation package.
The Base Pay Range For This Position Is $175,000 - $250,000 USD.
At WorldQuant, we are committed to providing candidates with all necessary information in compliance with pay transparency laws. If you believe any required details are missing from this job posting, please notify us at [email protected], and we will address your concerns promptly.
By submitting this application, you acknowledge and consent to terms of the WorldQuant Privacy Policy. The privacy policy offers an explanation of how and why your data will be collected, how it will be used and disclosed, how it will be retained and secured, and what legal rights are associated with that data (including the rights of access, correction, and deletion). The policy also describes legal and contractual limitations on these rights. The specific rights and obligations of individuals living and working in different areas may vary by jurisdiction.
#LI-RS1
By submitting this application, you acknowledge and consent to terms of the WorldQuant Privacy Policy. The privacy policy offers an explanation of how and why your data will be collected, how it will be used and disclosed, how it will be retained and secured, and what legal rights are associated with that data (including the rights of access, correction, and deletion). The policy also describes legal and contractual limitations on these rights. The specific rights and obligations of individuals living and working in different areas may vary by jurisdiction.
Copyright 2025 WorldQuant, LLC. All Rights Reserved.
WorldQuant is an equal opportunity employer and does not discriminate in hiring on the basis of race, color, creed, religion, sex, sexual orientation or preference, age, marital status, citizenship, national origin, disability, military status, genetic predisposition or carrier status, or any other protected characteristic as established by applicable law.
Data Scientist with SAS Enterprise Miner
Data engineer job in Jericho, NY
Title: Data Scientist with SAS Enterprise Miner 3- 5 years experience Location: Jericho, NY - source local first; ok to source national as long as the candidate is willing to relocate at their own expense Duration: 9+ months Visa Type: H1, US Citizen, GC
Interview: 1 or 2 rounds of phone followed by onsite. May accept video web conference for out of state candidates
Travel: no
Rate: open - market rates
Description:
Analytical Science team is looking for strong Data Scientists who will be primarily responsible for retraining various predictive models using new data source and techniques with the goal to outperformance the existing models.
Responsibilities:
* Retrain existing predictive models using new data source and possibly new advanced techniques
* Work comfortably with multi-terabyte and billion+ rows of data
* Perform rigorous model evaluation, design hypothesis tests, oversee test execution and result evaluation
Required skills/experience:
* Advanced degree in Machine Learning, Applied Math, Computer Science, Economics, Statistics or a related quantitative field
* At least 5 years recent experience (within past 3 years) building predictive models using SAS EG and SAS EM
* Proficiency in SAS Enterprise Miner/ Enterprise Guide/ Base SAS is critical
* Proficiency in Python and R in addition to SAS is a plus
* Solid understanding of advanced statistical concepts, especially related to modelling is a must
* Experience working with SQL Server, Google Cloud Platform/Google Big Query and Hadoop ecosystem
Preference will be given to candidates with:
* Practical experience in Neural Network, Random Forest, SVM and any other boosting techniques
* Background in collaborative filtering, data mining, machine learning, optimization or statistical theory
*
Please share the following candidate details with each resume submission:
* How many years SAS experience in past 3 years
* How many years SAS Enterprise Miner experience in past 3 years
* How many machine learning models built within past 3 years
OFSAA data modeler
Data engineer job in Norwalk, CT
At least 3-7 years of relevant experience in OFSAA for Banking Industry Strong Functional knowledge on Financial Risk management, MRM, Liquidity risk, Basel regulation Exposure to at least one risk management product in BFS space Ability to effectively liaise with multiple stakeholders both from business and IT.
Ability to analyze, define and document requirements related to data, workflow, risk process, interfaces with other systems
Ability to test the system thoroughly and ensure complete compliance with expected requirements.
Ability to manage change request and perform impact analysis on the existing requirement.
Ability to ensure software quality before release into production
Strong requirement collection, analysis and documentation skills
Exposure to use of requirement management tools
Strong knowledge on Data governance, data quality, Data Profiling & Data analysis related to Regulatory Reporting
OFSAA Implementation experience for regulatory reporting like Model risk management, Liquidity Risk, Basel etc.
OFSAA Data Model overview
Knowledge of technical tool stack: Informatica, Exadata, OBIEE etc.
Good exposure to Data level testing, Data validation, Quality & SIT
Working knowledge of BI applications
Desirable Skill Set :
ETL/Informatica
Qualifications
Bachelor or higher
Additional Information
Job Status: Permanent
Share the Profiles to *****************************
Contact:
************
Keep the subject line with Job Title and Location
Easy ApplyETL Development and Architecture-----------------Need GC and USC
Data engineer job in Greenwich, CT
USM Business Systems Inc. is a quickly developing worldwide System Integrator, Software and Product Development, IT Outsourcing and Technology assistance supplier headquartered in Chantilly, VA with off-shore delivery centers in India. We offer world-class ability in giving most astounding quality and administrations through industry best practices planned to convey remarkable worth to our customers.
Utilizing our industry knowledge, administration service offering expertise and innovation abilities, we distinguish new business and innovation slants and create answers for help customers around the globe, giving top of the line solid and practical IT benefits which are cost effective services.
Established in 1999, the organization has corner qualities in building and dealing with a Business Oriented IT environment with rich involvement in technology innovation, ERP and CRM counselling, Product Engineering, Business Intelligence, Data Management, SOA, BPM, Data Warehousing, SharePoint Consulting and IT Infrastructure. Our other offerings include modified solutions and administrations in ERP, CRM, Enterprise architecture, offshore advisory services ,e-commerce, Social , Mobile, Cloud, Analytics (SMAC) and DevOps.
USM, a US ensured Minority Business Enterprise (MBE) is perceived as one of the fastest developing IT Systems Integrator in the Washington, DC zone. Most as of late, USM was positioned #9 on the rundown of the Top administrations organizations in the DC Metro Area - Washington Business Journal (2011). We are a project-driven firm that reliably meets the IT needs of our State and Government customers through development and business keenness.
ETL Development and Architecture
Greenwich, CT
3 Years contract
Need GC and USC
Top Three Skills:
1. ETL Development and Architecture expertise leveraging SSIS on a SQL Server platform. (SSIS is a must have) Any knowledge of other ETL tools would be a plus as well.
2. Hedge Fund/Financials work experience, understanding of investment data and terminology from the Buy Side*.
3. Business facing with excellent communication and the ability to gather complex requirements from various teams across the organization.
4. Attribution Analysis and Asset Allocation expertise in prior projects will be a major plus within this organization.
Job Description:
The Information Technology Group of AQR Capital Management is seeking an experienced Data Warehouse and Business Intelligence professional to help the firm expand its use of data and BI tools to meet corporate goals and objectives. The individual will be providing optimal DW/ETL architecture and design for the enterprise DW environment. This includes gathering requirements, designing, working with the BI team and rolling solutions to end users. This individual will act as an advocate of DW/ETL best practices. Candidate must have effective communication skills and the ability to interact professionally within a diverse group of staff and user community. Candidate must be able to work productively in teams and to mentor or instruct others while effectively prioritizing their own tasks. Successful candidates will have great technology and problem solving skills, a solid work ethic, and an eagerness to learn in a highly intellectual, collaborative environment.
ROLE
Responsible for leading the strategic design and maintenance of Data Warehouse and ETL platforms.
Direct, organize, and lead projects in the implementation and use of ETL software tools and systems.
Design, code, test, and document all new or modified ETL processes, applications, and programs.
Manage the design, development, architecture, and delivery of all AQR internal and external reports.
Assist in the design of BI solutions to ensure interoperability with DW/ETL solutions.
Assist in analyzing user requirements and, based on findings, design ETL processes based on the functional specifications.
Adhere to best practices in change control and documentation.
REQUIREMENTS
8+ years of hands-on experience with DW/ETL processes, data architecture and business requirements gathering/analysis.
Expert knowledge of one ETL tool (Informatica, SSIS, Ab Initio etc.) - preferably SSIS
Sound knowledge of data warehouse methodologies.
Sound knowledge of data warehouse design methodologies (Kimball, Inmon etc.)
Strong SQL skills.
Good understanding of financial markets and products.
Direct experience in implementing enterprise DW/ETL processes and decision support.
Strong ability to analyze user requirements and build DW/ETL processes to specifications.
Direct hands-on experience with ad hoc query programs and reporting software.
OPTIONAL (Nice to Have)
Knowledge of data mining and predictive analysis
Understanding of .Net (C#) and/or Java
Understanding of fund structures and strategies.
Additional Information
If you are interested please share your resume to
preethib
@usmsystems(dot)com
or can directly call me on
************
C++ Market Data Engineer (USA)
Data engineer job in Stamford, CT
Trexquant is a growing systematic fund at the forefront of quantitative finance, with a core team of highly accomplished researchers and engineers. To keep pace with our expanding global trading operations, we are seeking a C++ Market Data Engineer to design and build ultra-low-latency feed handlers for premier vendor feeds and major exchange multicast feeds. This is a high-impact role that sits at the heart of Trexquant's trading platform; the quality, speed, and reliability of your code directly influence every strategy we run.
Responsibilities
* Design & implement high-performance feed handlers in modern C++ for equities, futures, and options across global venues (e.g., NYSE, CME, Refinitiv RTS, Bloomberg B-PIPE).
* Optimize for micro- and nanosecond latency using lock-free data structures, cache-friendly memory layouts, and kernel-bypass networking where appropriate.
* Build reusable libraries for message decoding, normalization, and publication to internal buses shared by research, simulation, and live trading systems.
* Collaborate with cross-functional teams to tune TCP/UDP multicast stacks, kernel parameters, and NIC settings for deterministic performance.
* Provide robust failover, gap-recovery, and replay mechanisms to guarantee data integrity under packet loss or venue outages.
* Instrument code paths with precision timestamping and performance metrics; drive continuous latency regression testing and capacity planning.
* Partner closely with quantitative researchers to understand downstream data requirements and to fine-tune delivery formats for both simulation and live trading.
* Produce clear architecture documents, operational run-books, and post-mortems; participate in a 24×7 follow-the-sun support rotation for mission-critical market-data services.
Programmer Analyst - Application Developer
Data engineer job in Saint James, NY
Title: Programmer Analyst - Application Developer Department: CPMP Information Technology Hiring Organization: StaffCo, Clinical Practice Management Plan (CPMP), Stony Brook Medicine
Schedule: Full-Time | Monday-Friday, 8:30 AM - 5:00 PM
Salary Range: $70,000 - $90,000
Budget ID: GNS2526-01
Reports To: Manager, Web and Application Development
Position Summary:
This full-time role is a key contributor to the digital transformation efforts of the Clinical Practice Management Plan (CPMP) and Stony Brook Medicine. The Programmer Analyst - Application Developer is responsible for designing, developing, and maintaining secure, scalable, and healthcare-focused applications that support clinical workflows, administrative operations, research initiatives, and patient engagement.
Working within a dynamic healthcare IT environment, the ideal candidate will bring full-stack development expertise and a strong understanding of healthcare technologies, including compliance (e.g., HIPAA), interoperability standards, and data security best practices. This position plays a vital role in enhancing operational efficiency and improving the digital experience for both providers and patients.
Responsibilities
Key Responsibilities
Collaborate with clinical and administrative stakeholders to develop application solutions that enhance patient care delivery and operational efficiency.
Support the full software development lifecycle (SDLC) for healthcare-related applications, from requirements gathering through deployment and maintenance.
Develop and maintain web applications using ASP.NET, .NET Core, C#, MVC, HTML, CSS, JavaScript, and SQL/T-SQL, with a focus on usability and accessibility.
Integrate third-party healthcare systems and APIs (e.g., financial platforms, scheduling tools) to ensure seamless data exchange and workflow automation.
Ensure applications meet HIPAA and other healthcare compliance standards, with attention to data privacy and security.
Utilize Microsoft Visual Studio for development and Azure DevOps, Git for version control.
Apply Agile methodologies to manage development tasks and collaborate effectively across multidisciplinary teams.
Use Azure DevOps and Git for source control and CI/CD pipelines.
Perform database development and optimization using MSSQL, Oracle, and Azure SQL, supporting clinical data integrity and reporting needs.
Manage hosting environments including IIS, Azure App, and other security protocols relevant to healthcare IT infrastructure.
Troubleshoot and resolve issues with a focus on minimizing disruption to clinical operations.
Stay current with healthcare IT trends and technologies, and proactively recommend improvements.
Qualifications
Required Qualifications
Bachelor's degree in computer science, Information Technology, or a related discipline, with a minimum of two years of hands-on experience in web and application development using the Microsoft technology stack.
In lieu of a bachelor's degree, candidates may qualify with four (4) years of relevant experience, or an associate degree combined with four (4) years of experience.
Preferred Qualifications:
2-4 years of experience developing or supporting healthcare IT applications.
Experience with Microsoft Azure, including deployment of healthcare web solutions, integration with Entra ID (formerly Azure Active Directory), Microsoft Graph, and AI-based development.
Proficiency with Microsoft Power Platform tools such as Power Apps and Power Automate for clinical workflow automation.
Knowledge of ETL processes and data integration strategies within healthcare environments.
Microsoft certifications in relevant technologies, including Azure development, AI, and database management.
Demonstrated commitment to ongoing professional development and continuous learning, particularly in healthcare IT.
Possesses Microsoft certifications in key areas such as Azure application development, AI integration, and database management, demonstrating advanced proficiency in cloud-based healthcare solutions.
The responsibilities and tasks outlined in this job description are not exhaustive and may change as determined by the needs of CPMP.
CPMP provides equal employment opportunities (EEO) to all employees and applicants for employment without regard to race, color, religion, creed, gender, national origin, age, disability, marital or veteran status, sexual orientation, gender identity or expression, or any other legally protected status. This policy applies to all terms and conditions of employment, including recruiting, hiring, placement, promotion, termination, layoff, recall and transfer, leaves of absence, compensation and training.
CPMP expressly prohibits any form of workplace harassment based on race, color, religion, creed, gender, national origin, age, disability, marital or veteran status, sexual orientation, gender identity, or any other legally protected status. Improper interference with the ability of CPMP's employees to perform their job duties may result in discipline up to and including discharge.
Auto-ApplySenior Data Engineer
Data engineer job in Farmingdale, NY
D'Addario & Company is the world's largest manufacturer and distributor of musical instrument accessories. As a U.S.-based manufacturing leader, we pride ourselves on high-automation machinery, cutting-edge technology, and a deep commitment to environmentally sustainable practices. Most importantly, we're proud of our diverse team of individuals who embody our core values-family, curiosity, passion, candor, and responsibility-and bring them to life every day.
D'Addario is seeking a Senior Data Engineer to help architect, build, and optimize the next generation of our global data infrastructure. In this role, you'll design and maintain production-grade data pipelines, support AI and machine learning initiatives, and serve as a technical mentor within a growing Business Intelligence team. You'll work closely with the Global Director of BI to deliver scalable solutions that power insights and innovation across the organization. This position is ideal for someone who thrives on solving complex data challenges, enjoys bringing structure to large datasets, and is passionate about enabling smarter decision-making through data.
This is a hybrid role and will require the candidate to work on-site in the Farmingdale office three days a week.
At D'Addario, we don't just offer a job-we offer a career with one of the most iconic names in the music industry. We're passionate about innovation, craftsmanship, and creating a workplace where diverse backgrounds, perspectives, and ideas thrive. We're eager to connect with individuals who bring fresh thinking and a collaborative spirit. If you're ready to make an impact, we'd love to hear how you'll add value to our team.
Some Perks & Benefits of Working at D'Addario:
Competitive compensation package
Health, vision, and dental insurance
12 weeks of fully paid parental leave
Fertility and family-building benefits
401(k) retirement plan with generous employer contributions
Career pathing and professional development via LinkedIn Learning
Paid Time Off (PTO) and flexible sick day policy
12 Paid Holidays
Life and AD&D Insurance
Enhanced Short-Term Disability Insurance
Employee Assistance Program (EAP)
Tuition Reimbursement
Discounts on D'Addario products and merchandise
Company jam nights, artist performances, holiday parties, and special events
A passionate, talented team that loves what they do!
Responsibilities
Build & Optimize Pipelines: Design, implement, and maintain robust, high-performance data pipelines to support analytical models within Microsoft Fabric and our data environment.
Data Integration: Connect and harmonize new data sources, including ERP, e-commerce platforms, and external APIs.
Mentorship & Standards: Guide junior BI team members, lead code reviews, and establish coding, documentation, and testing best practices.
AI/ML Enablement: Partner on machine learning and AI projects from proof-of-concept through deployment, supporting predictive and prescriptive analytics into production workflows.
Advanced Analytics Development: Team up with analysts to prepare data products and predictive models using Python, PySpark, and modern ML frameworks.
Collaboration: Work with stakeholders across Sales, Marketing, Operations, and Product to translate business requirements and align priorities.
Technical Leadership: Drive data engineering excellence through continuous improvement, quality assurance, and innovation in data architecture and governance.
Qualifications
5+ years of experience building and maintaining production-grade data pipelines.
Advanced programming skills in Python, PySpark, and SQL.
Strong background in data modeling and scalable analytics.
Experience deploying machine learning models and data products in production environments.
Solid understanding of cloud data platforms (Azure preferred).
Bachelor's degree in Computer Science, Engineering, Data Science, or equivalent experience.
Clear communicator with the ability to simplify complex technical concepts.
Proven leadership in mentoring and developing technical talent.
Highly organized, self-directed, and comfortable in fast-paced environments.
Passion for using data to drive innovation and business impact.
The base salary range for this role would be commensurate with experience: $140k to $165k per year
#LI-HYBRID
Auto-ApplyHadoop Developer
Data engineer job in Lake Success, NY
A big opportunity! Great pay! Love Hadoop? Do you have it in you to become a mentor? Then, my client requires YOUR mojo with them. This is an excellent opportunity for those aspiring for a growth oriented company. You will be the first person to work on Hadoop
in their New York office. You will be responsible for starting and educating new team members on Hadoop technologies, with a combination of hands on and training responsibilities.
Interesting?? Scroll down for more..
Job Description
Not an exhaustive set of responsibilities, but an
overview
:
You will be responsible for Architect, design and develop code that consistently adheres to functional programming principles.
You will design, develop, and maintain high volume data processing batch/streaming jobs using industry standard tools and frameworks in Hadoop ecosystem.
You will help my client continue as an industry leader by exploring new technologies,languages and techniques in the rapidly evolving world of high volume data processing.
You will collaborate with team members using Agile techniques, including: pair programming, test driven development (TDD), code reviews, and retrospectives.
If this has piqued your curiosity, please check if YOU have the following:
Experience in Hadoop Ecosystems (MapReduce/Spark/Hive/Oozie/scoop etc...)
Development experience in one of the following - Java, Python OR Scala.
Capable of working with and influencing large, diverse teams.
Experience working with large data sets and high volume data processing.
Here's what we can offer:
A competitive Base Salary of $140K with a Bonus of up to $10K and Benefits.
Note: We are looking for local candidates.
Additional Information
All your information will be kept confidential according to EEO guidelines.
Ping me at
******************** to know more.
Easy ApplyTech Lead, Data & Inference Engineer
Data engineer job in Greenwich, CT
Our Client
A fast moving and venture backed advertising technology startup based in San Francisco. They have raised twelve million dollars in funding and are transforming how business to business marketers reach their ideal customers. Their identity resolution technology blends business and consumer signals to convert static audience lists into high match and cross channel segments without the use of cookies. By transforming first party and third party data into precision targetable audiences across platforms such as Meta, Google and YouTube, they enable marketing teams to reach higher match rates, reduce wasted advertising spend and accelerate pipeline growth. With a strong understanding of how business buyers behave in channels that have traditionally been focused on business to consumer activity, they are redefining how business brands scale demand generation and account based efforts.
About Us
Catalyst Labs is a leading talent agency with a specialized vertical in Applied AI, Machine Learning, and Data Science. We stand out as an agency thats deeply embedded in our clients recruitment operations.
We collaborate directly with Founders, CTOs, and Heads of AI in those themes who are driving the next wave of applied intelligence from model optimization to productized AI workflows. We take pride in facilitating conversations that align with your technical expertise, creative problem-solving mindset, and long-term growth trajectory in the evolving world of intelligent systems.
Location: San Francisco
Work type: Full Time,
Compensation: above market base + bonus + equity
Roles & Responsibilities
Lead the design, development and scaling of an end to end data platform from ingestion to insights, ensuring that data is fast, reliable and ready for business use.
Build and maintain scalable batch and streaming pipelines, transforming diverse data sources and third party application programming interfaces into trusted and low latency systems.
Take full ownership of reliability, cost and service level objectives. This includes achieving ninety nine point nine percent uptime, maintaining minutes level latency and optimizing cost per terabyte. Conduct root cause analysis and provide long lasting solutions.
Operate inference pipelines that enhance and enrich data. This includes enrichment, scoring and quality assurance using large language models and retrieval augmented generation. Manage version control, caching and evaluation loops.
Work across teams to deliver data as a product through the creation of clear data contracts, ownership models, lifecycle processes and usage based decision making.
Guide architectural decisions across the data lake and the entire pipeline stack. Document lineage, trade offs and reversibility while making practical decisions on whether to build internally or buy externally.
Scale integration with application programming interfaces and internal services while ensuring data consistency, high data quality and support for both real time and batch oriented use cases.
Mentor engineers, review code and raise the overall technical standard across teams. Promote data driven best practices throughout the organization.
Qualifications
Bachelors or Masters degree in Computer Science, Computer Engineering, Electrical Engineering, or Mathematics.
Excellent written and verbal communication; proactive and collaborative mindset.
Comfortable in hybrid or distributed environments with strong ownership and accountability.
A founder-level bias for actionable to identify bottlenecks, automate workflows, and iterate rapidly based on measurable outcomes.
Demonstrated ability to teach, mentor, and document technical decisions and schemas clearly.
Core Experience
6 to 12 years of experience building and scaling production-grade data systems, with deep expertise in data architecture, modeling, and pipeline design.
Expert SQL (query optimization on large datasets) and Python skills.
Hands-on experience with distributed data technologies (Spark, Flink, Kafka) and modern orchestration tools (Airflow, Dagster, Prefect).
Familiarity with dbt, DuckDB, and the modern data stack; experience with IaC, CI/CD, and observability.
Exposure to Kubernetes and cloud infrastructure (AWS, GCP, or Azure).
Bonus: Strong Node.js skills for faster onboarding and system integration.
Previous experience at a high-growth startup (10 to 200 people) or early-stage environment with a strong product mindset.
Data Engineer
Data engineer job in New Haven, CT
Bexorg is transforming drug discovery by restoring molecular activity in postmortem human brains. Our groundbreaking BrainEx platform enables direct experimentation on functionally preserved human brain tissue, generating massive, high-fidelity molecular datasets that power AI-driven drug discovery for CNS diseases. We are seeking a Data Engineer to help harness this unprecedented data. In this onsite, mid-level role, you will design and optimize the pipelines and cloud infrastructure that turn terabytes of raw experimental data into actionable insights, driving our mission to revolutionize treatments for central nervous system disorders.
The Job:
Data Ingestion & Pipeline Management: Manage and optimize massive data ingestion pipelines from cutting-edge experimental devices, ensuring reliable, real-time capture of complex molecular data.
Cloud Data Architecture: Organize and structure large datasets in Google Cloud Platform, using tools like BigQuery and cloud storage to build a scalable data warehouse for fast querying and analysis of brain data.
Large-Scale Data Processing: Design and implement robust ETL/ELT processes to handle PB scale data, emphasizing speed, scalability, and data integrity at each step of the process.
Internal Data Services: Work closely with our software and analytics teams to expose processed data and insights to internal web applications. Build appropriate APIs or data access layers so that scientists and engineers can seamlessly visualize and interact with the data through our web platform.
Internal Experiment Services: Work with our life science teams to ensure data entry protocols for seamless metadata integration and association with experimental data
Infrastructure Innovation: Recommend and implement cloud infrastructure improvements (such as streaming technologies, distributed processing frameworks, and automation tools) that will future-proof our data pipeline. You will continually assess new technologies and best practices to increase throughput, reduce latency, and support our rapid growth in data volume.
Qualifications and Skills:
Experience with Google Cloud: Hands-on experience with Google Cloud services (especially BigQuery and related data tools) for managing and analyzing large datasets. You've designed or maintained data systems in a cloud environment and understand how to leverage GCP for big data workloads.
Data Engineering Background: 3+ years of experience in data engineering or a similar role. Proven ability to build and maintain data pipelines dealing with petabyte-scale data. Proficiency in programming (e.g., Python, Java, or Scala) and SQL for developing data processing jobs and queries.
Scalability & Performance Mindset: Familiarity with distributed systems or big data frameworks and a track record of optimizing data workflows for speed and scalability. You can architect solutions that handle exponential data growth without sacrificing performance.
Biology Domain Insight: Exposure to biology or experience working with scientific data (e.g. genomics, bioinformatics, neuroscience) is a strong plus. While deep domain expertise isn't required, you should be excited to learn about our experimental data and comfortable discussing requirements with biologists.
Problem-Solving & Collaboration: Excellent problem-solving skills, attention to detail, and a proactive attitude in tackling technical challenges. Ability to work closely with cross-functional teams (scientists, software engineers, data scientists) and communicate complex data systems in clear, approachable terms.
Passion for the Mission: A strong desire to apply your skills to transform drug discovery. You are inspired by Bexorg's mission and eager to build the data backbone of a platform that could unlock new therapies for CNS diseases.
Local to New Haven, CT preferred. No relocation offered for this position.
Bexorg is an equal opportunity employer. We strive to create a supportive and inclusive workplace where contributions are valued and celebrated, and our employees thrive by being themselves and are inspired to do their best work. We seek applicants of all backgrounds and identities, across race, color, ethnicity, national origin or ancestry, citizenship, religion, sex, sexual orientation, gender identity or expression, veteran status, marital status, pregnancy or parental status, or disability. Applicants will not be discriminated against based on these or other protected categories or social identities. Bexorg will also consider for employment qualified applicants with criminal histories in a manner consistent with applicable federal, state and local law.
Data Platform Engineer (USA)
Data engineer job in Stamford, CT
Trexquant is a growing systematic fund at the forefront of quantitative finance, with a core team of highly accomplished researchers and engineers. To keep pace with our expanding global trading operations, we are seeking a highly motivated and technically rigorous Data Platform Engineer to help modernize our foundational data infrastructure. As a Data Platform Engineer, you will be at the center of building the systems that ensure the quality, reliability, and discoverability of mission-critical data. Your work will directly impact the data operators and downstream consumers by creating robust tools, monitoring, and workflows that ensure accuracy, validity, and timeliness of data across the firm.
Responsibilities
* Architect and maintain core components of the Data Platform with a strong focus on reliability and scalability.
* Build and maintain tools to manage data feeds, monitor validity, and ensure data timeliness.
* Design and implement event-based data orchestration pipelines.
* Evaluate and integrate data quality and observability tools via POCs and MVPs.
* Stand up a data catalog system to improve data discoverability and lineage tracking.
* Collaborate closely with infrastructure teams to support operational excellence and platform uptime.
* Write and maintain data quality checks to validate real-time and batch data.
* Validate incoming real-time data using custom Python-based validators.
* Ensure low-level data correctness and integrity, especially in high-precision environments.
* Build robust and extensible systems that will be used by data operators to ensure the health of our data ecosystem.
* Own the foundational systems used by analysts and engineers alike to trust and explore our datasets.
OFSAA Data Architect
Data engineer job in Norwalk, CT
Mandatory Technical Skills : Strong in data warehousing concepts and dimension modeling - Min 6 Years Exp. Experience in OFSAA Data Modeling etc - Min 3 years Translate business requirements into (OFSAA) designs and map data elements from models to the OFSAA data model.
Strong troubleshooting skills
Hands-on experience with extracting, loading of data from source systems into the OFSAA model.
Data modeling (star / 3NF / cube), ETL design and build.
Extensive experience in OFSAA Infrastructure, OFSAA Data Model, & Erwin Data Modeler.
Desirable Technical Skills : OBIEE Analytics and BI - ETL Knowledge
Mandatory Functional Skills :
Ability to co-ordinate with multiple technical teams, Business users and Customer
Strong communication
Strong troubleshooting skills
Should have strong understanding on OFSAA LRM, Basel, OBIEE Analytics."
Desirable Functional Skills : Banking and finance service Industry
Qualifications
Bachelor or higher
Additional Information
Job Status: Permanent
Share the Profiles to *****************************
Contact:
************
Keep the subject line with Job Title and Location
Easy ApplyData Platform Engineer (USA)
Data engineer job in Stamford, CT
Job Description
Trexquant is a growing systematic fund at the forefront of quantitative finance, with a core team of highly accomplished researchers and engineers. To keep pace with our expanding global trading operations, we are seeking a highly motivated and technically rigorous Data Platform Engineer to help modernize our foundational data infrastructure. As a Data Platform Engineer, you will be at the center of building the systems that ensure the quality, reliability, and discoverability of mission-critical data. Your work will directly impact the data operators and downstream consumers by creating robust tools, monitoring, and workflows that ensure accuracy, validity, and timeliness of data across the firm.
Responsibilities
Architect and maintain core components of the Data Platform with a strong focus on reliability and scalability.
Build and maintain tools to manage data feeds, monitor validity, and ensure data timeliness.
Design and implement event-based data orchestration pipelines.
Evaluate and integrate data quality and observability tools via POCs and MVPs.
Stand up a data catalog system to improve data discoverability and lineage tracking.
Collaborate closely with infrastructure teams to support operational excellence and platform uptime.
Write and maintain data quality checks to validate real-time and batch data.
Validate incoming real-time data using custom Python-based validators.
Ensure low-level data correctness and integrity, especially in high-precision environments.
Build robust and extensible systems that will be used by data operators to ensure the health of our data ecosystem.
Own the foundational systems used by analysts and engineers alike to trust and explore our datasets.
Requirements
A Bachelor's degree in Computer Science or a related field; advanced degree preferred.
3+ years of hands-on experience with Python in a data engineering or backend development context
Experience with distributed data systems (e.g., Spark, Kafka, Airflow).
Proven experience running POCs, evaluating data quality and data platform tools.
Demonstrated interest and experience in low-level data reliability, correctness, and observability.
Familiarity with systems-level thinking and the principles of data operations in production.
Background in high-performance computing or real-time data processing is a plus.
Prior experience in a quantitative or financial setting is highly desirable.
Benefits
Competitive salary, plus bonus based on individual and company performance.
Collaborative, casual, and friendly work environment while solving the hardest problems in the financial markets.
PPO Health, dental, and vision insurance premiums fully covered for you and your dependents.
Pre-Tax Commuter Benefits - making your commute smoother.
Trexquant is an Equal Opportunity Employer