Today
Secret
Early Career (2+ yrs experience)
$150,000 and above
IT - Security
Remote/Hybrid•Arlington, VA (On-Site/Office)
ALTA IT Services is looking for a Data Engineer to work for a large government contractor in Arlington, VA. This is an exciting opportunity to join a great team and bring your skills to the table.
The ideal candidate MUST be well versed in Python. We are looking for someone who is a "jack of all trades" and can do data engineering, analysis and science. This is a multifaceted role and will require being comfortable pivoting and assisting in different areas when necessary.
Duration: Six-month temp to perm
Pay Rate: 150k-165k depending on experience
Clearance: Active DoD Secret - would love higher
Work Model: Hybrid - 3 days on-site in Arlington VA
This role requires deep understanding of data architecture, data warehousing, ETL processes and big data technologies. The Data Engineer will work closely with data scientists, analyst, and other stakeholders to ensure the seamless flow and integrity of data across the organization.
Responsibilities:
Build, maintain, and optimize data pipelines for extracting, transforming, and loading (ETL) data from various sources into our data warehouse.
Integrate large, complex data sets that meet functional and non-functional business requirements.
Design and implement scalable and reliable data warehousing solutions.
Extract data from various sources, transform it into a usable format, and load it into Azure data storage solutions such as Azure Data Lake Storage, Azure SQL Database, or Azure Synapse Analytics
Develop and maintain data pipelines using Azure Data Factory and other relevant Azure technologies
Manage and optimize databases, ensuring data quality, integrity, and security.
Automate manual processes, optimize data delivery, and re-design infrastructure for greater scalability.
Collaborate with data scientists, analysts, and other stakeholders to understand data needs and deliver high-quality data solutions.
Document data flows, processes, and system architecture to ensure clarity and knowledge sharing within the team.
Required:
An active DoD Secret clearance.
BA/BS in Computer Science, Information Technology, Data Analytics, or related field.
5+ years of related experience.
Strong experience in data extract transform load (ETL) processes using Azure services, such as Azure Data Factory, Azure Databricks, Azure Data Lake Storage, Azure SQL Database, or Azure Synapse Analytics.
4+ years’ experience and skill writing coding languages (such as SQL, Python, R, SAS, and Java Scripts).
2+ years’ experience working with projects involving machine learning, natural language processing, robotics process automation, artificial intelligence, text and/or data mining, as well as statistical and mathematical methods.
Experience with using SQL to conduct complex database queries.
Experience with AI/ML.
Desired:
2+ years of experience working within the DoD or IC Community.
Master's degree in Mathematics, Statistics, CS, or Data Analytics.
Clearance: TS with SCI Eligibility.
The project is specifically focusing on managed identities and cyber threats. You will be dealing with very large data sets of audit logs that will need to be cleaned and processed to use for data modeling.
The ideal candidate MUST be well versed in Python. We are looking for someone who is a "jack of all trades" and can do data engineering, analysis and science. This is a multifaceted role and will require being comfortable pivoting and assisting in different areas when necessary.
Duration: Six-month temp to perm
Pay Rate: 150k-165k depending on experience
Clearance: Active DoD Secret - would love higher
Work Model: Hybrid - 3 days on-site in Arlington VA
This role requires deep understanding of data architecture, data warehousing, ETL processes and big data technologies. The Data Engineer will work closely with data scientists, analyst, and other stakeholders to ensure the seamless flow and integrity of data across the organization.
Responsibilities:
Build, maintain, and optimize data pipelines for extracting, transforming, and loading (ETL) data from various sources into our data warehouse.
Integrate large, complex data sets that meet functional and non-functional business requirements.
Design and implement scalable and reliable data warehousing solutions.
Extract data from various sources, transform it into a usable format, and load it into Azure data storage solutions such as Azure Data Lake Storage, Azure SQL Database, or Azure Synapse Analytics
Develop and maintain data pipelines using Azure Data Factory and other relevant Azure technologies
Manage and optimize databases, ensuring data quality, integrity, and security.
Automate manual processes, optimize data delivery, and re-design infrastructure for greater scalability.
Collaborate with data scientists, analysts, and other stakeholders to understand data needs and deliver high-quality data solutions.
Document data flows, processes, and system architecture to ensure clarity and knowledge sharing within the team.
Required:
An active DoD Secret clearance.
BA/BS in Computer Science, Information Technology, Data Analytics, or related field.
5+ years of related experience.
Strong experience in data extract transform load (ETL) processes using Azure services, such as Azure Data Factory, Azure Databricks, Azure Data Lake Storage, Azure SQL Database, or Azure Synapse Analytics.
4+ years’ experience and skill writing coding languages (such as SQL, Python, R, SAS, and Java Scripts).
2+ years’ experience working with projects involving machine learning, natural language processing, robotics process automation, artificial intelligence, text and/or data mining, as well as statistical and mathematical methods.
Experience with using SQL to conduct complex database queries.
Experience with AI/ML.
Desired:
2+ years of experience working within the DoD or IC Community.
Master's degree in Mathematics, Statistics, CS, or Data Analytics.
Clearance: TS with SCI Eligibility.
The project is specifically focusing on managed identities and cyber threats. You will be dealing with very large data sets of audit logs that will need to be cleaned and processed to use for data modeling.
group id: COMPHLP