Job ID: 2411792
Location: MCLEAN, VA, United States
Date Posted: Oct 3, 2024
Category: Information Technology
Subcategory: Data Scientist
Schedule: Full-time
Shift: Day Job
Travel: Yes, 10 % of the Time
Minimum Clearance Required: TS/SCI with Poly
Clearance Level Must Be Able to Obtain: None
Potential for Remote Work: No Remote
Benefits: Click here
Description
Support a program which leverages integrated discrete technologies to support massive data processing, storage, modeling, and analytics over several thousand unique data sources, to perform threat identification and analysis, as well as support efforts to meet tactical and strategic goals. The components that comprise the data platform capability – the technologies and systems, data, data processing and modeling, and use of the data via data science and querying of the data corpus and model(s) to derive insights. The data platform capability serves as the backbone for other capabilities (e.g., web applications), to accelerate its operations.
Develop and use data processing technologies (e.g., Python, SPARK, Java, SQL, Jenkins, PyPi, Terraform, Cloudera, ElasticSearch, Pentaho, Apache NiFi, Apache Hop) to perform data processing, and to develop, validate, and use methodologies to support analytic requirements in Clustered Computing environments.
a. Support downstream systems and capabilities of external Customer organizations dependent on the data platform via various approaches, to include application programming interfaces (APIs).
b. On an ongoing basis, develop integration plans that capitalize on new data processing, modeling, and storage technologies including the cloud environments.
c. Evaluate data collections to assess the potential value-add to the Customers data platform to recommend to the Customer.
d. Generate assessments about data, support activities to perform data acquisition and engineering, and enable the processing of data so it is integrated into data platform systems for maximum value.
e. Perform and support data modeling and engineering activities for integration of new data into the data platforms data corpus, refining existing models and intermediate models to address deficiencies and defects, and with Customer oversight, creating new models and data feeds to support existing and new analytic methodologies.
Qualifications
Required Skills
1. Python
2. SPARK
3. Java
4. SQL
5. Jenkins
6. PyPi
7. Terraform and Cloudera
8. ElasticSearch
9. Pentaho
10. Apache NiFi
11. Apache Hop
12. Perform data processing, and to develop, validate, and use methodologies to support analytic requirements in Clustered Computing environments.
13. Perform and support data modeling and engineering activities for integration of new data into the data platforms data corpus, refining existing models and intermediate models to address deficiencies and defects, and with Customer oversight, creating new models and data feeds to support existing and new analytic methodologies.
Desired Skills
1. Demonstrated experience using Enterprise Control Language (ECL) and the Lexis-Nexis High Performance Cluster Computing (HPCC) platform.
2. Demonstrated experience performing All-Source data analysis to perform analytic support to the Sponsor.
3. Demonstrated experience developing custom algorithms to support analytic requirements against massive data stores supporting the Sponsor.
4. Demonstrated experience directly supporting the Sponsor performing technical analysis support using massive data processing systems.
5. Demonstrated experience writing cables.
6. Demonstrated experience planning and coordinating program activities such as installation and upgrading of hardware and software, utilization of cloud services, programming, or systems design development, modification of IT networks, or implementation of Internet and intranet sites.
7. Demonstrated experience deploying web applications to a cloud managed environment to include DevOps and security configuration management.
8. Demonstrated experience developing, implementing, and maintaining cloud infrastructure services such as EC2, ELB, RDS, S3, and VPC.
9. Demonstrated experience planning, coordinating, and executing the required activities to support documentation to meet the Sponsor’s data compliance requirements (e.g., legal, data policy)
10. Degree(s)
a. Undergraduate degree in mathematics, computer science, engineering, or similar scientific or technical discipline.
i. Graduate degree in computer science, information systems, engineering, or another scientific or technical discipline.
b. Degree or equivalent in CS, MIS, Economics, Physics, Genetics, or Engineering related field, especially Supercomputing-related.
SAIC accepts applications on an ongoing basis and there is no deadline.
Overview
SAIC® is a premier Fortune 500® technology integrator focused on advancing the power of technology and innovation to serve and protect our world. Our robust portfolio of offerings across the defense, space, civilian and intelligence markets includes secure high-end solutions in mission IT, enterprise IT, engineering services and professional services. We integrate emerging technology, rapidly and securely, into mission critical operations that modernize and enable critical national imperatives.
We are approximately 24,000 strong; driven by mission, united by purpose, and inspired by opportunities. SAIC is an Equal Opportunity Employer, fostering a culture of diversity, equity and inclusion, which is core to our values and important to attract and retain exceptional talent. Headquartered in Reston, Virginia, SAIC has annual revenues of approximately $7.4 billion. For more information, visit saic.com. For ongoing news, please visit our newsroom.