Software Engineer 2 – Linux/Bash/Python/Apache Spark/NumPy/Git
Description
Captivation Software is looking for a mid level software engineer who shall be responsible for identifying ways to create consistent and repeatable capabilities including transforming raw, complex, and often unstructured data into clean, reliable, and high-quality analysis-ready datasets that support advanced analytics, predictive modeling, and data-driven decision-making across the organization.
Requirements
Security Clearance:
- Must currently hold a Top Secret/SCI U.S. Government security clearance with a favorable Polygraph, therefore all candidates must be a U.S. citizen
Minimum Qualifications:
- Master’s degree in Computer Science or related discipline from an accredited college or university, plus three (3) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity.
- Bachelor’s degree in Computer Science or related discipline from an accredited college or university, plus five (5) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity
- Seven (7) years of experience as a SWE, in programs and contracts of similar scope, type, and complexity.
Required Skills:
- Experience using the Linux CLI and Linux tools
- Experience developing Bash scripts to automate manual processes
- Recent software development experience using Python
- Familiar with Distributed Big Data processing engines including Apache Spark
- Experience using Jupyter Notebook
- Experience with data wrangling and preprocessing tools such as pandas and NumPy
- Experience working with structured, semi-structured, and unstructured data
- Familiarity with data quality concepts, data validation, and anomaly detection
- Experience with Git Source Control System
Desired Skills:
- Familiar with Apache Airflow (DAG design, scheduling, operators, sensors) to orchestrate, schedule, and monitor complex workflows
- Familiar with SQL technologies such as MySQL, MariaDB, and PostgreSQL for querying, joining, and aggregating large datasets
- Familiar with HPC Job Scheduling tools including Slurm
- Experience using the Atlassian Tool Suite (JIRA, Confluence)
This position is open for direct hires only. We will not consider candidates from third party staffing/recruiting firms.