Senior Data Scientist
Senior data scientist job in McLean, VA
Purpose:
As a Data Scientist, you will play a key role in delivering impactful, data-driven solutions for our strategic enterprise clients. This role also offers the opportunity to shape and grow Infocepts' Data Science & AI practice, contributing to high-impact AI/ML initiatives, crafting data-driven narratives for stakeholders, and applying advanced techniques to solve complex business problems from strategy to execution.
Key Result Areas and Activities:
Design, build, and deploy AI/ML solutions using modern cloud and data platforms.
Lead data science projects across industries, ensuring alignment with business goals.
Apply supervised, unsupervised, deep learning, and Generative AI (e.g., LLMs, agentic workflows) techniques to address client use cases.
Collaborate with data engineering teams to optimize model pipelines using Delta Lake and Spark.
Communicate findings effectively through data visualizations and stakeholder presentations.
Drive adoption of MLOps practices for scalable and reliable model deployment.
Contribute to the evolution of Infocepts' Data Science & AI offerings through innovation and knowledge sharing.
Roles & Responsibilities
Essential Skills
5+ years of experience in applied AI and machine/deep learning.
Hands-on experience with Databricks, MLflow, PySpark, and Spark MLlib.
Proficiency in Python and SQL for model development and data manipulation.
Strong understanding of supervised and unsupervised learning, deep learning, and Generative AI.
Familiarity with cloud platforms: AWS, Azure, and GCP.
Solid foundation in advanced statistical methods and probabilistic analysis.
Ability to lead end-to-end AI/ML projects, including design, development, and stakeholder management.
Experience with visualization tools like Tableau, Power BI, or similar.
Familiarity with ML workflow orchestration and MLOps practices.
Desirable Skills
Experience with LLMs (Large Language Models) and agentic AI workflows.
Familiarity with modern data platforms like Snowflake.
Exposure to real-time data processing in cloud-native environments.
Contributions to open-source AI projects or publications in data science communities.
Qualifications
Bachelor's or Master's degree in Computer Science, Data Science, Machine Learning, Statistics, or a related field.
Certifications in cloud platforms (AWS, Azure, GCP) or Databricks are a plus.
Qualities:
Able to consult, write, and present persuasively
Able to work in a self-organized and cross-functional team
Able to iterate based on new information, peer reviews, and feedback
Able to work seamlessly with clients across multiple geographies
Research focused mindset
Excellent analytical, presentation, reporting, documentation and interactive skills
"Infocepts is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to sex, gender identity, sexual orientation, race, color, religion, national origin, disability, protected Veteran status, age, or any other characteristic protected by law."
Data Scientist - ML, Python
Senior data scientist job in McLean, VA
10+years of experience required in Information Technology.
Python Programming: At least 5 years of hands-on experience with Python, particularly in
frameworks like FastAPI, Django, Flask, and experience using AI frameworks.
• Access Control Expertise: Strong understanding of access control models such as Role-Based
Access Control (RBAC) and Attribute-Based Access Control (ABAC).
• API and Connector Development: Experience in developing API connectors using Python for
extracting and managing access control data from platforms like Azure, SharePoint, Java, .NET,
WordPress, etc.
• AI and Machine Learning: Hands-on experience integrating AI into applications for automating
tasks such as access control reviews and identifying anomalies
• Cloud and Microsoft Technologies: Proficiency with Azure services, Microsoft Graph API, and
experience integrating Python applications with Azure for access control reviews and reporting.
• Reporting and Visualization: Experience using reporting libraries in Python (Pandas, Matplotlib,
Plotly, Dash) to build dashboards and reports related to security and access control metrics.
• Communication Skills: Ability to collaborate with various stakeholders, explain complex
technical solutions, and deliver high-quality solutions on time.
• PlainID: Experience or familiarity with PlainID platforms for identity and access management.
• Azure OpenAI: Familiarity with Azure OpenAI technologies and their application in access
control and security workflows.
• Power BI: Experience with Microsoft Power BI for data visualization and reporting.
• Agile Methodologies: Experience working in Agile environments and familiarity with Scrum
methodologies for delivering security solutions.
Senior Data Scientist
Senior data scientist job in McLean, VA
Locals to Only# In- Person Interview
Job Title: Data Scientist Specialist
We are seeking a highly experienced Principal Gen AI Scientist with a strong focus on Generative AI (GenAI) to lead the design and development of cutting-edge AI Agents, Agentic Workflows and Gen AI Applications that solve complex business problems. This role requires advanced proficiency in Prompt Engineering, Large Language Models (LLMs), RAG, Graph RAG, MCP, A2A, multi-modal AI, Gen AI Patterns, Evaluation Frameworks, Guardrails, data curation, and AWS cloud deployments. You will serve as a hands-on Gen AI (data) scientist and critical thought leader, working alongside full stack developers, UX designers, product managers and data engineers to shape and implement enterprise-grade Gen AI solutions.
Responsibilities:
Architect and implement scalable AI Agents, Agentic Workflows and GenAI applications to address diverse and complex business use cases.
Develop, fine-tune, and optimize lightweight LLMs; lead the evaluation and adaptation of models such as Claude (Anthropic), Azure OpenAI, and open-source alternatives.
Design and deploy Retrieval-Augmented Generation (RAG) and Graph RAG systems using vector databases and knowledge bases.
Curate enterprise data using connectors integrated with AWS Bedrock's Knowledge Base/Elastic.
Implement solutions leveraging MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication.
Build and maintain Jupyter-based notebooks using platforms like AWS SageMaker and MLFlow/Kubeflow on Kubernetes (EKS).
Collaborate with cross-functional teams of UI and microservice engineers, designers, and data engineers to build full-stack Gen AI experiences.
Integrate GenAI solutions with enterprise platforms via API-based methods and GenAI standardized patterns.
Establish and enforce validation procedures with Evaluation Frameworks, bias mitigation, safety protocols, and guardrails for production-ready deployment.
Design & build robust ingestion pipelines that extract, chunk, enrich, and anonymize data from PDFs, video, and audio sources for use in LLM-powered workflows-leveraging best practices like semantic chunking and privacy controls.
Orchestrate multimodal pipelines** using scalable frameworks (e.g., Apache Spark, PySpark) for automated ETL/ELT workflows appropriate for unstructured media.
Implement embeddings drives-map media content to vector representations using embedding models, and integrate with vector stores (AWS Knowledge Base/Elastic/Mongo Atlas) to support RAG architectures.
Qualifications:
experience in AI/ML, with applied GenAI or LLM-based solutions.
Deep expertise in prompt engineering, fine-tuning, RAG, GraphRAG, vector databases (e.g., AWS Knowledge Base / Elastic), and multi-modal models.
Proven experience with cloud-native AI development (AWS SageMaker, Amazon Bedrock, MLFlow on EKS).
Strong programming skills in Python and ML libraries (Transformers, LangChain, etc.).
Deep understanding of Gen AI system patterns and architectural best practices, Evaluation Frameworks.
Demonstrated ability to work in cross-functional agile teams.
Data Scientist
Senior data scientist job in McLean, VA
Kavaliro is seeking a Data Scientist to provide highly technical and in-depth data engineering support. The candidate MUST have experience designing and building data infrastructure, developing data pipelines, transforming and preparing data, ensuring data quality and security, and monitoring and optimizing systems. The candidate MUST have extensive experience with Python and AWS. Experience with SQL, multi-data source queries with database technologies (PostgreSQL, MySQL, RDS, etc.), NiFi, Git, Elasticsearch, Kibana, Jupyter Notebooks, NLP, AI, and any data visualization tools (Tableau, Kibana, Qlik, etc.) are desired.
Required Skills and Demonstrated Experience
Demonstrated experience with data engineering, to include designing and building data infrastructure, developing data pipelines, transforming/preparing data, ensuring data quality and security, and monitoring/optimizing systems.
Demonstrated experience with data management and integration, including designing and perating robust data layers for application development across local and cloud or web data sources.
Demonstrated work experience programming with Python
Demonstrated experience building scalable ETL and ELT workflows for reporting and analytics.
Demonstrated experience with general Linux computing and advanced bash scripting
Demonstrated experience with SQL.
Demonstrated experience constructing complex multi-data source queries with database technologies such as PostgreSQL, MySQL, Neo4J or RDS
Demonstrated experience processing data sources containing structured or unstructured data
Demonstrated experience developing data pipelines with NiFi to bring data into a central environment
Demonstrated experience delivering results to stakeholders through written documentation and oral briefings
Demonstrated experience using code repositories such as Git
Demonstrated experience using Elastic and Kibana
Demonstrated experience working with multiple stakeholders
Demonstrated experience documenting such artifacts as code, Python packages and methodologies
Demonstrated experience using Jupyter Notebooks
Demonstrated experience with machine learning techniques including natural language processing
Demonstrated experience explaining complex technical issues to more junior data scientists, in graphical, verbal, or written formats
Demonstrated experience developing tested, reusable and reproducible work
Work or educational background in one or more of the following areas: mathematics, statistics, hard sciences (e.g. Physics, Computational Biology, Astronomy, Neuroscience, etc.) computer science, data science, or business analytics
Desired Skills and Demonstrated Experience
Demonstrated experience with cloud services, such as AWS, as well as cloud data technologies and architecture.
Demonstrated experience using big data processing tools such as Apache Spark or Trino
Demonstrated experience with machine learning algorithms
Demonstrated experience with using container frameworks such as Docker or Kubernetes
Demonstrated experience with using data visualizations tools such as Tableau, Kibana or Apache Superset
Demonstrated experience creating learning objectives and creating teaching curriculum in technical or scientific fields
Location:
McLean, Virginia
This position is onsite and there is no remote availability.
Clearance:
TS/SCI with Full Scope Polygraph
Applicant MUST hold a permanent U.S. citizenship for this position in accordance with government contract requirements.
Kavaliro provides Equal Employment Opportunities to all employees and applicants. All qualified applicants will receive consideration for employment without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state, or local laws. Kavaliro is committed to the full inclusion of all qualified individuals. In keeping with our commitment, Kavaliro will take the steps to assure that people with disabilities are provided reasonable accommodations. Accordingly, if reasonable accommodation is required to fully participate in the job application or interview process, to perform the essential functions of the position, and/or to receive all other benefits and privileges of employment, please respond to this posting to connect with a company representative.
Principal Data Scientist with Gen AI
Senior data scientist job in McLean, VA
Title: Principal Data Scientist with Gen AI
Contract: W2
Exp: 10+
Duration: Long Term
Interview Mode: In-Person interview
Call Notes:
Looking for a Principal Data Scientist with strong focus on Generative AI (GenAI) with expertise in Machine Learning transitioned into GenAI. Need someone with good experience in RAG, Python- Jupyter, other Software knowledge, using agents in workflows, strong understanding of data.
Someone with advanced proficiency in Prompt Engineering, Large Language Models (LLMs), RAG, Graph RAG, MCP, A2A, multi-modal AI, Gen AI Patterns, Evaluation Frameworks, Guardrails, data curation, and AWS cloud deployments.
Highly preferred for someone who can built AI agent, MCP, A2A, Graph Rag, deployed Gen AI applications to production.
Top Skills:
Machine Learning & Deep Learning - Required
GenAI - Required
Python - Required
Rag and/or Graph Rag - Required
MCP (Model Context Protocol) and A2A (Agent-to-Agent) is highly preferred
Job Description:
We are seeking a highly experienced **Principal Gen AI Scientist** with a strong focus on **Generative AI (GenAI)** to lead the design and development of cutting-edge AI Agents, Agentic Workflows and Gen AI Applications that solve complex business problems. This role requires advanced proficiency in Prompt Engineering, Large Language Models (LLMs), RAG, Graph RAG, MCP, A2A, multi-modal AI, Gen AI Patterns, Evaluation Frameworks, Guardrails, data curation, and AWS cloud deployments. You will serve as a hands-on Gen AI (data) scientist and critical thought leader, working alongside full stack developers, UX designers, product managers and data engineers to shape and implement enterprise-grade Gen AI solutions.
Key Responsibilities:
* Architect and implement scalable AI Agents, Agentic Workflows and GenAI applications to address diverse and complex business use cases.
* Develop, fine-tune, and optimize lightweight LLMs; lead the evaluation and adaptation of models such as Claude (Anthropic), Azure OpenAI, and open-source alternatives.
* Design and deploy Retrieval-Augmented Generation (RAG) and Graph RAG systems using vector databases and knowledge bases.
* Curate enterprise data using connectors integrated with AWS Bedrock's Knowledge Base/Elastic
* Implement solutions leveraging MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication.
* Build and maintain Jupyter-based notebooks using platforms like SageMaker and MLFlow/Kubeflow on Kubernetes (EKS).
* Collaborate with cross-functional teams of UI and microservice engineers, designers, and data engineers to build full-stack Gen AI experiences.
* Integrate GenAI solutions with enterprise platforms via API-based methods and GenAI standardized patterns.
* Establish and enforce validation procedures with Evaluation Frameworks, bias mitigation, safety protocols, and guardrails for production-ready deployment.
* Design & build robust ingestion pipelines that extract, chunk, enrich, and anonymize data from PDFs, video, and audio sources for use in LLM-powered workflows-leveraging best practices like semantic chunking and privacy controls
* Orchestrate multimodal pipelines** using scalable frameworks (e.g., Apache Spark, PySpark) for automated ETL/ELT workflows appropriate for unstructured media
* Implement embeddings drives-map media content to vector representations using embedding models, and integrate with vector stores (AWS KnowledgeBase/Elastic/Mongo Atlas) to support RAG architectures
Required Qualifications:**
* 10+ years of experience in AI/ML, with 3+ years in applied GenAI or LLM-based solutions.
* Deep expertise in prompt engineering, fine-tuning, RAG, GraphRAG, vector databases (e.g., AWS KnowledgeBase / Elastic), and multi-modal models.
* Proven experience with cloud-native AI development (AWS SageMaker, Bedrock, MLFlow on EKS).
* Strong programming skills in Python and ML libraries (Transformers, LangChain, etc.).
* Deep understanding of Gen AI system patterns and architectural best practices, Evaluation Frameworks
* Demonstrated ability to work in cross-functional agile teams.
* Need Github Code Repository Link for each candidate. Please thoroughly vet the candidates.
**Preferred Qualifications:**
* Published contributions or patents in AI/ML/LLM domains.
* Hands-on experience with enterprise AI governance and ethical deployment frameworks.
* Familiarity with CI/CD practices for ML Ops and scalable inference APIs.
Data scientist
Senior data scientist job in Reston, VA
Job title: Data scientist
Fulltime
About Smart IT Frame:
At Smart IT Frame, we connect top talent with leading organizations across the USA. With over a decade of staffing excellence, we specialize in IT, healthcare, and professional roles, empowering both clients and candidates to grow together.
Note:
• In- person interview
Must Have;
• Data science
• Python
• SQL
• ML/Ops
• Risk Modelling
📩 Apply today or share profiles at ****************************
Data Scientist
Senior data scientist job in Columbia, MD
Data Scientist - Transit Data Focus_Columbia, MD (On-site / hybrid)_Contract (6 Months)
Data Scientist - Transit Data Focus
Employment type: Contract
Duration: 6 Months
Justification: To manage and analyze customer databases, AVA (automated voice announcement), and schedule data for predictive maintenance and service planning.
Experience Level: 3-5 years
Job Responsibilities:
Collect, process, and analyze transit-related datasets including customer databases, AVA (automated voice announcement) logs, real-time vehicle data, and schedule data.
Develop predictive models and data-driven insights to support maintenance forecasting, service planning, and operational optimization.
Design and implement data pipelines to integrate, clean, and transform large, heterogeneous transit data sources.
Perform statistical analysis and machine learning to identify patterns, trends, and anomalies relevant to transit service performance and reliability.
Collaborate with transit planners, maintenance teams, and IT staff to translate data insights into actionable business strategies.
Monitor data quality and integrity; implement data validation and cleansing processes.
Technical Skills & Qualifications:
Bachelor's or Master's degree in Data Science, Statistics, Computer Science, Transportation Engineering, or a related quantitative field.
3-5 years of experience working as a data scientist or data analyst, preferably in a transit, transportation, or public sector environment.
Strong proficiency in Python or R for data analysis, statistical modeling, and machine learning.
Experience with SQL for database querying, manipulation, and data extraction.
Familiarity with transit data standards such as GTFS, AVL/CAD, APC (Automated Passenger Counters), and AVA systems.
Experience with data visualization tools such as Power BI, or equivalent.
Data Scientist Specialist
Senior data scientist job in McLean, VA
Job Title: Data Scientist Specialist
Duration: 45 Minutes|120 Minutes
Interview Type: MS Teams - Video Mandatory| 2nd round on-site
Call notes:
Flexibility in hands on experience
Not required 10 years of experience
ML and Gen AI
2 years of Gen AI experience
Gen AI development
This role comes between software engineer and Data Scientist
Can see Data Engineer
BS/MS in Ai or Data Science preferred
PHD degree not needed
Senior Data Scientist
Senior data scientist job in McLean, VA
We are seeking a highly experienced **Principal Gen AI Scientist** with a strong focus on **Generative AI (GenAI)** to lead the design and development of cutting-edge AI Agents, Agentic Workflows and Gen AI Applications that solve complex business problems. This role requires advanced proficiency in Prompt Engineering, Large Language Models (LLMs), RAG, Graph RAG, MCP, A2A, multi-modal AI, Gen AI Patterns, Evaluation Frameworks, Guardrails, data curation, and AWS cloud deployments. You will serve as a hands-on Gen AI (data) scientist and critical thought leader, working alongside full stack developers, UX designers, product managers and data engineers to shape and implement enterprise-grade Gen AI solutions.
Key Responsibilities:
* Architect and implement scalable AI Agents, Agentic Workflows and GenAI applications to address diverse and complex business use cases.
* Develop, fine-tune, and optimize lightweight LLMs; lead the evaluation and adaptation of models such as Claude (Anthropic), Azure OpenAI, and open-source alternatives.
* Design and deploy Retrieval-Augmented Generation (RAG) and Graph RAG systems using vector databases and knowledge bases.
* Curate enterprise data using connectors integrated with AWS Bedrock's Knowledge Base/Elastic
* Implement solutions leveraging MCP (Model Context Protocol) and A2A (Agent-to-Agent) communication.
* Build and maintain Jupyter-based notebooks using platforms like SageMaker and MLFlow/Kubeflow on Kubernetes (EKS).
* Collaborate with cross-functional teams of UI and microservice engineers, designers, and data engineers to build full-stack Gen AI experiences.
* Integrate GenAI solutions with enterprise platforms via API-based methods and GenAI standardized patterns.
* Establish and enforce validation procedures with Evaluation Frameworks, bias mitigation, safety protocols, and guardrails for production-ready deployment.
* Design & build robust ingestion pipelines that extract, chunk, enrich, and anonymize data from PDFs, video, and audio sources for use in LLM-powered workflows-leveraging best practices like semantic chunking and privacy controls
* Orchestrate multimodal pipelines** using scalable frameworks (e.g., Apache Spark, PySpark) for automated ETL/ELT workflows appropriate for unstructured media
* Implement embeddings drives-map media content to vector representations using embedding models, and integrate with vector stores (AWS KnowledgeBase/Elastic/Mongo Atlas) to support RAG architectures
**Required Qualifications:**
* 10+ years of experience in AI/ML, with 3+ years in applied GenAI or LLM-based solutions.
* Deep expertise in prompt engineering, fine-tuning, RAG, GraphRAG, vector databases (e.g., AWS KnowledgeBase / Elastic), and multi-modal models.
* Proven experience with cloud-native AI development (AWS SageMaker, Bedrock, MLFlow on EKS).
* Strong programming skills in Python and ML libraries (Transformers, LangChain, etc.).
* Deep understanding of Gen AI system patterns and architectural best practices, Evaluation Frameworks
* Demonstrated ability to work in cross-functional agile teams.
* Need Github Code Repository Link for each candidate. Please thoroughly vet the candidates.
**Preferred Qualifications:**
* Published contributions or patents in AI/ML/LLM domains.
* Hands-on experience with enterprise AI governance and ethical deployment frameworks.
* Familiarity with CI/CD practices for ML Ops and scalable inference APIs.
#LI-CGTS
#TS-2942
Data Modeler
Senior data scientist job in Washington, DC
Job Title: Developer Premium I
Duration: 7 Months with long term extension
Hybrid Onsite: 4 days per week from Day 1
Job Requirement:
Strong expertise in Data Architecture & Date model design.
MS Azure (core experiment)
Experience with SAP ECC preferred
SAFE agile certification is a plus
Ability to work flexibility including off hours to support critical IT task & migration activities.
Educational Qualifications and Experience:
Bachelor's degree in Computer Science, Information Systems or in a related area of expertise.
Required number of years of proven experience in the specific technology/toolset as per Experience Matrix below for each Level.
Essential Job Functions:
Take functional specs and produce high quality technical specs
Take technical specs and produce completed and well tested programs which meet user satisfaction and acceptance, and precisely reflect the requirements - business logic, performance, and usability requirements
Conduct/attend requirements definition meetings with end-users and document system/business requirements
Conduct Peer Review on Code and Test Cases, prepared by other team members, to assess quality and compliance with coding standards
As required for the role, perform end-user demos of proposed solution and finished product, provide end user training and provide support for user acceptance testing
As required for the role, troubleshoot production support issues and find appropriate solutions within defined SLA to ensure minimal disruption to business operations
Ensure that Bank policies, procedures, and standards are factored into project design and development
As required for the role, install new release, and participate in upgrade activities
As required for the role, perform integration between systems that are on prem and also on the cloud and third-party vendors
As required for the role, collaborate with different teams within the organization for infrastructure, integration, database administration support
Adhere to project schedules and report progress regularly
Prepare weekly status reports and participate in status meetings and highlight issues and constraints that would impact timely delivery of work program items
Find the appropriate tools to implement the project
Maintain knowledge of current industry standards and practices
As needed, interact and collaborate with Enterprise Architects (EA), Office of Information Security (OIS) to obtain approvals and accreditations
“Mindlance is an Equal Opportunity Employer and does not discriminate in employment on the basis of - Minority/Gender/Disability/Religion/LGBTQI/Age/Veterans.”
Senior Data Modeler
Senior data scientist job in Baltimore, MD
6 Month Contract (Possible Extension or Contract-to-Hire)
* Must be able to work on a W-2
* Must be able to work EST hours
Job Description
Seeking a seasoned Data Modeler to architect and operationalize scalable, governed data structures that transform raw, disparate data into trusted, analytics-ready assets. This role demands mastery of data modeling principles, normalization, and SQL-based transformations within modern cloud ecosystems (e.g., Snowflake). As a senior individual contributor, you will set modeling standards, influence cross-functional teams, and ensure data integrity, consistency, and performance across the enterprise.
• Ability to collaborate closely with technical and business leaders to understand data requirements and deliver impactful solutions.
• Ability to design data quality, data lineage, and data governance practices across the organization.
• Ability to advocate for advanced data engineering tools, platforms, and technologies.
• Ability to transfer knowledge and experience to junior staff.
Key Responsibilities
Data Architecture & Modeling
• Design and maintain conceptual, logical, and physical data models that enable high-performance analytics and transactional workloads.
• Normalize and optimize complex, raw, and semi-structured data into relational structures aligned with business and analytical requirements.
• Own metadata and lineage: document entity relationships, definitions, and transformation logic to ensure transparency and reusability.
Collaboration & Influence
• Partner strategically with data engineers, analysts, and business stakeholders to translate requirements into scalable models and influence upstream design decisions.
• Serve as SME for data modeling best practices; evangelize standards and mentor peers to elevate modeling maturity across teams.
Implementation & Governance
• Implement models in Snowflake and other cloud platforms, ensuring compliance with performance, security, and governance frameworks.
• Champion automation: leverage dbt, Python, or orchestration tools to streamline model deployment and enforce version control.
• Conduct proactive data profiling and quality checks to validate accuracy and integrity before production rollout.
Innovation & Continuous Improvement
• Lead POCs for modern modeling approaches-including semantic layers, ML-ready schemas, and hybrid architectures-and institutionalize successful patterns.
• Continuously refine modeling standards, naming conventions, and governance policies to support evolving business needs.
Mandatory Skills
• 5+ years of experience in data modeling or database design, with proven ability to normalize complex datasets into optimized relational models.
• Advanced SQL skills and strong understanding of ETL/ELT processes.
• Hands-on experience with modern cloud data platforms (Snowflake preferred).
Education
• Bachelor's or Master's degree in Computer Science, Engineering, or a related field.
Desired Skills
• Familiarity with data governance and cataloging tools (e.g., Collibra, Alation).
• Experience with dbt or scripting (Python) for automated transformations and CI/CD workflows.
• Strong communication skills to translate technical concepts into business language and influence stakeholders.
You Will Receive The Following Benefits:
Medical & Dental Insurance
Health Savings Account (HSA)
401(k)
Paid Sick Time Leave
Pre-tax Commuter Benefit
Applicants must be currently authorized to work in the United States on a full-time basis now and in the future.
Motion Recruitment provides IT Staffing Solutions (Contract, Contract-to-Hire, and Direct Hire) in major North American markets. Our unique expertise in today's highest-demand tech skill sets, paired with our deep networks and knowledge of our local technology markets, results in an exemplary track record with candidates and clients.
Azure Data Modeler
Senior data scientist job in Washington, DC
Azure Data Modeler - Budget Transformation Project
Our client is embarking on a major budget transformation initiative and is seeking an experienced Azure Data Modeler to support data architecture, modeling, and migration activities. This role will play a critical part in designing and optimizing data structures as the organization transitions to SAP. Experience with SAP is preferred, but strong ERP data experience in any platform is also valuable.
Responsibilities
Design, develop, and optimize data models within the Microsoft Azure environment.
Support data architecture needs across the budget transformation program.
Partner with cross-functional stakeholders to enable the transition to SAP (or other ERP systems).
Participate in data migration planning, execution, and validation efforts.
Work collaboratively within SAFe Agile teams and support sprint activities.
Provide off-hours support as needed for critical tasks and migration windows.
Engage onsite in Washington, DC up to three days per week.
Required Qualifications
Strong hands-on expertise in data architecture and data model design.
Proven experience working with Microsoft Azure (core requirement).
Ability to work flexibly, including occasional off-hours support.
Ability to be onsite in Washington, DC as needed (up to 3 days/week).
Preferred Qualifications
Experience with SAP ECC or exposure to SAP implementations.
Experience with other major ERP systems (Oracle, Workday, etc.).
SAFe Agile certification.
Dexian stands at the forefront of Talent + Technology solutions with a presence spanning more than 70 locations worldwide and a team exceeding 10,000 professionals. As one of the largest technology and professional staffing companies and one of the largest minority-owned staffing companies in the United States, Dexian combines over 30 years of industry expertise with cutting-edge technologies to deliver comprehensive global services and support.
Dexian connects the right talent and the right technology with the right organizations to deliver trajectory-changing results that help everyone achieve their ambitions and goals. To learn more, please visit ********************
Dexian is an Equal Opportunity Employer that recruits and hires qualified candidates without regard to race, religion, sex, sexual orientation, gender identity, age, national origin, ancestry, citizenship, disability, or veteran status.
Data Engineer / Big data Engineer
Senior data scientist job in McLean, VA
Immediate need for a talented Data Engineer / Big data Engineer. This is a 12 months contract opportunity with long-term potential and is located in Mclean, VA(Hybrid). Please review the job description below and contact me ASAP if you are interested.
Job ID: 25-93504
Pay Range: $70 - $75/hour. Employee benefits include, but are not limited to, health insurance (medical, dental, vision), 401(k) plan, and paid sick leave (depending on work location).
Key Responsibilities:
Design, develop, and maintain data pipelines leveraging Python, Spark/PySpark, and cloud-native services.
Build and optimize data workflows, ETL processes, and transformations for large-scale structured and semi-structured datasets.
Write advanced and efficient SQL queries against Snowflake, including joins, window functions, and performance tuning.
Develop backend and automation tools using Golang and/or Python as needed.
Implement scalable, secure, and high-quality data solutions across AWS services such as S3, Lambda, Glue, Step Functions, EMR, and CloudWatch.
Troubleshoot complex production data issues, including pipeline failures, data quality gaps, and cloud environment challenges.
Perform root-cause analysis and implement automation to prevent recurring issues.
Collaborate with data scientists, analysts, platform engineers, and product teams to enable reliable, high-quality data access.
Ensure compliance with enterprise governance, data quality, and cloud security standards.
Participate in Agile ceremonies, code reviews, and DevOps practices to ensure high engineering quality.
Key Requirements and Technology Experience:
Proficiency in Python with experience building scalable data pipelines or ETL processes.
Strong hands-on experience with Spark/PySpark for distributed data processing.
Experience writing complex SQL queries (Snowflake preferred), including optimization and performance tuning.
Working knowledge of AWS cloud services used in data engineering (S3, Glue, Lambda, EMR, Step Functions, CloudWatch, IAM).
Experience with Golang for scripting, backend services, or performance-critical processes.
Strong debugging, troubleshooting, and analytical skills across cloud and data ecosystems.
Familiarity with CI/CD workflows, Git, and automated testing.
Our client is a leading Banking and Financial Industry, and we are currently interviewing to fill this and other similar contract positions. If you are interested in this position, please apply online for immediate consideration.
Pyramid Consulting, Inc. provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.
By applying to our jobs you agree to receive calls, AI-generated calls, text messages, or emails from Pyramid Consulting, Inc. and its affiliates, and contracted partners. Frequency varies for text messages. Message and data rates may apply. Carriers are not liable for delayed or undelivered messages. You can reply STOP to cancel and HELP for help. You can access our privacy policy here.
Data Engineer
Senior data scientist job in Falls Church, VA
*** W2 Contract Only - No C2C - No 3rd Parties ***
The Ash Group is hiring a new Programmer Analyst Principal (Data Engineer) for our client (a global leader providing advanced systems and support in defense, aerospace, and security) based in Falls Church, VA.
In this role, you'll be designing, implementing, and optimizing large-scale data systems and ETL pipelines, with a strong focus on using Amazon Redshift and AWS services to ensure data quality and integrity for complex defense programs.
Compensation, Benefits, and Role Info
Competitive pay rate of $65 per hour.
Medical, dental, vision, direct primary care benefits, and, after six months of employment, a 4% matched 401(k) plan with immediate 100% vesting.
Type: 12-month contract with potential extension or conversion.
Location: On-site in Falls Church, VA.
What You'll Be Doing
Design and implement large-scale ETL data pipelines using AWS Glue and Python/PySpark to ingest, transform, and load data from various sources.
Build and maintain robust data warehouses, focusing on Amazon Redshift, including data modeling and governance.
Write and optimize complex, highly-performant SQL queries across large datasets (Redshift, Oracle, SQL Server).
Collaborate with cross-functional teams (data scientists, analysts) to understand requirements and deliver end-to-end data solutions.
Troubleshoot, optimize performance, and resolve data-related issues like pipeline failures and data quality bottlenecks.
What We're Looking For
8+ years of hands-on experience in data engineering, focusing on designing and implementing large-scale data systems.
5+ years of experience in building production-level ETL pipelines using AWS Glue and Python/PySpark.
Deep proficiency in SQL, including query optimization, indexing, and performance tuning across data warehouses like Amazon Redshift.
Strong understanding of database design principles, data modeling (star/snowflake schemas), and data governance.
Experience with data processing/orchestration frameworks such as Apache Airflow, Apache Kafka, or Fivetran.
If you're a seasoned data engineering professional passionate about building scalable data solutions and driving innovation in cloud-based environments, we want to hear from you. This is an exciting opportunity to work on cutting-edge technologies, collaborate with cross-functional teams, and make a meaningful impact on data-driven decision-making. Apply now to be part of a forward-thinking organization where your expertise will shape the future of our data infrastructure.
#DataEngineer #DataEngineering #AWSEngineer #Redshift #ETL #PySpark #DataPipeline #Westminster #ColoradoJobs #Contract
Senior Data Scientist
Senior data scientist job in Springfield, VA
Job Title: Senior Data Scientist
Workplace: On-site
Clearance Required: TS/SCI with willingness to sit for CI Polygraph
We are seeking a talented and motivated Senior Data Scientist with expertise in statistical analysis, artificial intelligence and machine learning, to join our team. In this role, you will be responsible for developing advanced algorithms and data-driven solutions that address complex challenges for our customers within the U.S. Intelligence Community. In addition, were looking for someone interested in growing as a leader.
Roles and Responsibilities:
Support client strategies through the creation of automated collection models, dynamic analytic models, workflow automations, and any other automation processes and products as assigned.
Refine, enhance, and improve the operational performance of any automated solution through the evaluation of performance data, regular customer interaction, and a standardized maintenance cycle.
Apply data science and visual programming tradecraft to support and streamline analysis tasks as identified by stakeholders and the government.
Enhance technical solutions to problems related to IC intelligence integration, automated collections, tipping and cueing, information sharing, and visualization.
Conduct extensive collections and analytic modeling, data processing, data mining, and visualization. Conduct gap analysis on existing technologies and processes.
Provide communication to customers on the progress for projects, processes, and emerging technologies as they become available.
Clearly communicate data-driven findings and automation to technical and non-technical audiences.
Conduct customer elicitation to identify processing problems due to procedures, tools, and services; work arounds that may need a permanent solution; and gaps in tools and technology.
Train and integrate new tools, processes, and capabilities to be used in collection orchestration.
Provide support for emerging requirements as assigned by the government.
Position Requirements:
Required Clearance: TS/SCI with willingness to sit for CI Polygraph
Required Education: Bachelor's degree in a relevant field
Required Skills / Experience:
6+ years of Data Science experience
3+ years of experience with a programming language, preferably Python
3+ years of experience with a querying language, preferably SQL
Experience in AI/ML/DS
Excellent communication skills with experience briefing/presenting to senior leaders
Ability to manage multiple simultaneous technical projects
Preferred Skills and Qualifications:
Master's degree in a relevant field with 4+ years of experience
Database management
A continuous learning mindset
Team leadership experience
Good interpersonal skills - ability to build strong relationships with customers, teaming partners, and co-workers
About Elder Research, Inc
People Centered. Data Driven
Elder Research is a fast growing consulting firm specializing in predictive analytics. Being in the data mining business almost 30 years, we pride ourselves in our ability to find creative, cutting edge solutions to real-world problems. We work hard to provide the best value to our clients and allow each person to contribute their ideas and put their skills to use immediately.
Our team members are passionate, curious, life-long learners. We value humility, servant-leadership, teamwork, and integrity. We seek to serve our clients and our teammates to the best of our abilities. In keeping with our entrepreneurial spirit, we want candidates who are self-motivated with an innate curiosity and strong team work.
Elder Research believes in continuous learning and community - each week the entire company attends a Tech Talk and each office location provides lunch. Elder Research provides a supportive work environment with established parental, bereavement, and PTO policies. By prioritizing a healthy work-life balance - with reasonable hours, solid pay, low travel, and extremely flexible time off - Elder Research enables and encourages its employees to serve others and enjoy their lives.
Elder Research, Inc. is an Equal Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, or protected veteran status and will not be discriminated against on the basis of disability.
Elder Research is a Government contractor and many of our positions require US Citizenship.
Senior Data Scientist (Model Audit)
Senior data scientist job in Vienna, VA
Navy Federal Credit Union currently does not provide sponsorship for this role. Applicants must be authorized to work in the United States without the need for current or future sponsorship. Navy Federal's Internal Audit team is in the midst of an exciting transformational journey to become a best in-class Audit function! It is our vision to be a preferred advisor to the business by building and cultivating trust through the consistent execution of high-quality and risk-focused audit and advisory work. We're focused on implementing efficient processes, maximizing our use of technology, integrating data analytics into everything we do, and investing in our biggest asset, our people. If this sounds like the type of team you'd like to be a part of, then we want to learn more about you!
Support Internal Audit's evaluation of model risk frameworks and their ability to mitigate risk for individual models and in the aggregate. Conduct reviews of complex data science and machine learning models across the organization through audit engagements and technical reviews of models. Serve as a subject matter expert to Internal Audit staff, senior management and business partners on end-to-end life-cycle management of model development, implementation, ongoing monitoring, regulatory expectations, and industry modeling practices. Conduct work assignments of increasing complexity, under moderate supervision with some latitude for independent judgment. Intermediate professional within field; requires moderate skill set and proficiency in discipline.
* Responsible for assessing model governance processes including model inherent risk ratings, model validations, materiality of model changes, and other reviews of models as part of model -related governance audits.
* Leverages specialized technical approaches to develop, implement, and execute model testing plans for models that require expert model risk knowledge and skills.
* Executes independent internal audit reviews of models to ensure the independent model audit process compliant with the written risk and compliance policies and procedures at a subject matter expert level.
* Evaluates model development, implementation, and use as it pertains to elements of model conceptual soundness, model assumptions, modeling methodology, model limitations, data quality, ongoing monitoring, and other more complex elements of the model development process.
* Identifies meaningful insights from large data and metadata sources
* Test hypotheses/models, analyze, and interpret results
* Exercise sound judgment and discretion within defined procedures and practices
* Develop and code moderately complex software programs, algorithms, and automated processes
* Use modeling and trend analysis to analyze data and provide insights
* Develop understanding of best practices and ethical AI
* Transform data into charts, tables, or format that aids effective decision making
* Build working relationships with team members and subject matter experts
* Lead small projects and initiatives
* Utilize effective written and verbal communication to document and present findings of analyses to a diverse audience of stakeholders
* 5-7 years of experience in data analysis, statistical modeling or regression analysis, including language models, LLMs and/or Generative AI technologies
* Basic understanding of business and operating environment
* Statistics
* Programming, data modeling, simulation, and advanced mathematics
* SQL, R, Python, Hadoop, SAS, SPSS, Scala, AWS, Databricks
* Model lifecycle execution including model development, validation, and governance
* Technical writing
* Data storytelling and technical presentation skills
* Research Skills
* Interpersonal Skills
* Working knowledge of procedures, instructions, and validation techniques
* Model Development
* Communication
* Critical Thinking
* Collaborate and Build Relationships
* Initiative with sound judgement
* Technical (Big Data Analysis, Coding, Project Management, Technical Writing, etc.)
* Sound Judgment
* Problem Solving (Responds as problems and issues are identified)
* Master's Degree in Data Science, Statistics, Mathematics, Computers Science, Engineering, or another quantitative or related field
Desired Qualifications:
* Prior experience with enterprise risk management, model risk management, or internal audit
* Deep knowledge and experience with SR 11-7 and/or ASOP 56
* Understanding of various models and modeling practices used in credit risk management, fraud detection, BSA/AML, operations, treasury & finance, marketing models, etc.
* Knowledge of one or more regulations such as CECL, CCAR, Anti-Money Laundering, ECOA, FCRA, etc.
* Advanced knowledge of Navy Federal Credit Union instructions, standards, and procedures
Hours: Monday - Friday, 8:00AM - 4:30PM
Locations: 820 Follin Lane, Vienna, VA 22180 | 5510 Heritage Oaks Drive, Pensacola, FL 32526
Auto-ApplyData Scientist
Senior data scientist job in Reston, VA
• Collect, clean, and preprocess large datasets from multiple sources.
• Apply statistical analysis and machine learning techniques to solve business problems.
• Build predictive models and algorithms to optimize processes and improve outcomes.
• Develop dashboards and visualizations to communicate insights effectively.
• Collaborate with cross-functional teams (Product, Engineering, Risk, Marketing) to identify opportunities for leveraging data.
• Ensure data integrity, security, and compliance with organizational standards.
• Stay current with emerging technologies and best practices in data science and AI.
________________________________________
Required Qualifications
• Bachelor's or Master's degree in Data Science, Computer Science, Statistics, Mathematics, or related field.
• Strong proficiency in Python, R, SQL, and experience with data manipulation libraries (e.g., Pandas, NumPy).
• Hands-on experience with machine learning frameworks (e.g., scikit-learn, TensorFlow, PyTorch).
• Solid understanding of statistical modeling, hypothesis testing, and data visualization.
• Experience with big data platforms (e.g., Spark, Hadoop) and cloud environments (AWS, Azure, GCP).
• Excellent problem-solving skills and ability to communicate complex concepts clearly.
________________________________________
Preferred Qualifications
• Experience in risk modeling, financial services, or product analytics.
• Knowledge of MLOps and deploying models in production.
• Familiarity with data governance and compliance frameworks.
________________________________________
Soft Skills
• Strong analytical thinking and attention to detail.
• Ability to work independently and in a team environment.
• Effective communication and stakeholder management skills.
#LI-CGTS
#TS-0455
AWS Data Engineer
Senior data scientist job in McLean, VA
Responsibilities:
Design, build, and maintain scalable data pipelines using AWS Glue and Databricks.
Develop and optimize ETL/ELT processes using PySpark and Python.
Collaborate with data scientists, analysts, and stakeholders to enable efficient data access and transformation.
Implement and maintain data lake and warehouse solutions on AWS (S3, Glue Catalog, Redshift, Athena, etc.).
Ensure data quality, consistency, and reliability across systems.
Optimize performance of large-scale distributed data processing workflows.
Develop automation scripts and frameworks for data ingestion, transformation, and validation.
Follow best practices for data governance, security, and compliance.
Required Skills & Experience:
5-8 years of hands-on experience in Data Engineering.
Strong proficiency in Python and PySpark for data processing and transformation.
Expertise in AWS services - particularly Glue, S3, Lambda, Redshift, and Athena.
Hands-on experience with Databricks for building and managing data pipelines.
Experience working with large-scale data systems and optimizing performance.
Solid understanding of data modeling, data lake architecture, and ETL design principles.
Strong problem-solving skills and ability to work independently in a fast-paced environment.
“Mindlance is an Equal Opportunity Employer and does not discriminate in employment on the basis of - Minority/Gender/Disability/Religion/LGBTQI/Age/Veterans.”
Senior Data Engineer.
Senior data scientist job in McLean, VA
Immediate need for a talented Senior Data Engineer. This is a 06+months contract opportunity with long-term potential and is located in Mclean, VA(Remote). Please review the job description below and contact me ASAP if you are interested.
Job ID: 25-84666
Pay Range: $64 - $68/hour. Employee benefits include, but are not limited to, health insurance (medical, dental, vision), 401(k) plan, and paid sick leave (depending on work location).
Key Responsibilities:
Demonstrate ability in implementing data warehouse solutions using modern data platforms such as Client, Databricks or Redshift.
Build data integration solutions between transaction systems and analytics platforms.
Expand data integration solutions to ingest data from internal and external sources and to further transform as per the business consumption needs.
Develop tasks for a multitude of data patterns, e.g., real-time data integration, advanced analytics, machine learning, BI and reporting.
Fundamental understanding of building of data products by data enrichment and ML.
Act as a team player and share knowledge with the existing team members.
Key Requirements and Technology Experience:
Key skills; Python, AWS, SNOWFLAKE
Bachelor's degree in computer science or a related field.
Minimum 5 years of experience in building data driven solutions.
At least 3 years of experience working with AWS services.
Applicants must be authorized to work in the US without requiring employer sponsorship currently or in the future. U.S. FinTech does not offer H-1B sponsorship for this position.
Expertise in real-time data solutions, good-to-have knowledge of streams processing, Message Oriented Platforms and ETL/ELT Tools.
Strong scripting experience using Python and SQL.
Working knowledge of foundational AWS compute, storage, networking and IAM.
Understanding of Gen AI models, prompt engineering, RAG, fine tuning and pre-tuning will be a plus.
Solid scripting experience in AWS using Lambda functions.
Knowledge of CloudFormation template preferred.
Hands-on experience with popular cloud-based data warehouse platforms such as Redshift and Client.
Experience in building data pipelines with related understanding of data ingestion, transformation of structured, semi-structured and unstructured data across cloud services.
Knowledge and understanding of data standards and principles to drive best practices around data management activities and solutions.
Experience with one or more data integration tools such as Attunity (Qlik), AWS Glue ETL, Talend, Kafka etc.
Strong understanding of data security - authorization, authentication, encryption, and network security.
Hands on experience in using and extending machine learning framework and libraries, e.g, scikit-learn, PyTorch, TensorFlow, XGBoost etc. preferred.
Experience with AWS SageMaker family of services or similar tools to develop machine learning models preferred.
Strong written and verbal communication skills to facilitate meetings and workshops to collect data, functional and technology requirements, document processes, data flows, gap analysis, and associated data to support data management/governance related efforts.
Acts with integrity and proactively seeks ways to ensure compliance with regulations, policies, and procedures.
Demonstrated ability to be self-directed with excellent organization, analytical and interpersonal skills, and consistently meet or exceed deadline deliverables.
Strong understanding of the importance and benefits of good data quality, and the ability to champion results across functions.
Our client is a leading Financial Industry, and we are currently interviewing to fill this and other similar contract positions. If you are interested in this position, please apply online for immediate consideration.
Pyramid Consulting, Inc. provides equal employment opportunities to all employees and applicants for employment and prohibits discrimination and harassment of any type without regard to race, color, religion, age, sex, national origin, disability status, genetics, protected veteran status, sexual orientation, gender identity or expression, or any other characteristic protected by federal, state or local laws.
By applying to our jobs you agree to receive calls, AI-generated calls, text messages, or emails from Pyramid Consulting, Inc. and its affiliates, and contracted partners. Frequency varies for text messages. Message and data rates may apply. Carriers are not liable for delayed or undelivered messages. You can reply STOP to cancel and HELP for help. You can access our privacy policy here.
Cloud Data Engineer- Databricks
Senior data scientist job in McLean, VA
Purpose:
We are seeking a highly skilled Cloud Data Engineer with deep expertise in Databricks and modern cloud platforms such as AWS, Azure, or GCP. This role is ideal for professionals who are passionate about building next-generation data platforms, optimizing complex data workflows, and enabling advanced analytics and AI in cloud-native environments. You'll have the opportunity to work with Fortune-500 organizations in data and analytics, helping them unlock the full potential of their data through innovative, scalable solutions.
Key Result Areas and Activities:
Design and implement robust, scalable data engineering solutions.
Build and optimize data pipelines using Databricks, including serverless capabilities, Unity Catalog, and Mosaic AI.
Collaborate with analytics and AI teams to enable real-time and batch data workflows.
Support and improve cloud-native data platforms (AWS, Azure, GCP).
Ensure adherence to best practices in data modeling, warehousing, and governance.
Contribute to automation of data workflows using CI/CD, DevOps, or DataOps practices.
Implement and maintain workflow orchestration tools like Apache Airflow and dbt.
Roles & Responsibilities
Essential Skills
4+ years of experience in data engineering with a focus on scalable solutions.
Strong hands-on experience with Databricks in a cloud environment.
Proficiency in Spark and Python for data processing.
Solid understanding of data modeling, data warehousing, and architecture principles.
Experience working with at least one major cloud provider (AWS, Azure, or GCP).
Familiarity with CI/CD pipelines and data workflow automation.
Desirable Skills
Direct experience with Unity Catalog and Mosaic AI within Databricks.
Working knowledge of DevOps/DataOps principles in a data engineering context.
Exposure to Apache Airflow, dbt, and modern data orchestration frameworks.
Qualifications
Bachelor's or Master's degree in Computer Science, Data Engineering, Information Systems, or a related field.
Relevant certifications in cloud platforms (AWS/Azure/GCP) or Databricks are a plus.
Qualities:
Able to consult, write, and present persuasively
Able to work in a self-organized and cross-functional team
Able to iterate based on new information, peer reviews, and feedback
Able to work seamlessly with clients across multiple geographies
Research focused mindset
Excellent analytical, presentation, reporting, documentation and interactive skills
"Infocepts is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to sex, gender identity, sexual orientation, race, color, religion, national origin, disability, protected Veteran status, age, or any other characteristic protected by law."