DataBricks/Spark Engineer

Job Locations US-VA-McLean
ID
2021-1102
Category
Computer/Data Science
Position Type
Regular Full-Time
Clearance Level
Able to obtain

Overview

BigBear.ai is seeking a DataBricks/Spark Engineer to join our Engineering Team. You will apply your skills in data science, data analytics, and data warehousing to pioneer cloud and data services with our clients in this position. You will use cloud and data technologies to deliver exciting IT & data-driven capabilities. This position will support NuWave Solutions, a BigBear.ai company. 

What you will do

  • Engineer, build and maintain scalable, automated data pipelines. Support critical data pipelines with a highly scalable distributed architecture - including data ingestion (streaming, events, and batch), data integration, data curation
  • Build cloud data solutions and provide domain perspective on storage, big data platform services, serverless architecture, RDBMS, DW/DM, NoSQL databases, and security.
  • Deploy, automate, maintain and manage cloud-based production systems to ensure the availability, performance, scalability, and security of production systems
  • Conduct full technical discovery, identifying pain points, business, and technical requirements, “as is” and “to be” scenarios
  • Design and arrangement of scalable, highly attainable, and fault-tolerant big data ecosystems on a cloud platform
  • Ownership and responsibility for end-to-end design and development, testing, the release of critical components data lake solution using DataBricks
  • Understand and implement best practices in managing data, including master data, reference data, metadata, data quality, and lineage
  • Support regular ad-hoc data querying and analysis to understand customer behaviors

What you need to have

  • Must be US Citizen with an ability to obtain and maintain US government clearance suitability
  • BA or BS degree in CS, Information Systems, Information Systems
  • At least 3 years of DataBricks/Sparks Engineering experience
  • Experience with developing and deploying ETL pipelines using Apache Spark
  • Experience with distributed computing, such as Apache Spark and/or Databricks
  • Experience in SQL and/or Spark SQL
  • Experience with code versioning tools and a command of configuration management concepts and tools, CI/CD including DevOps.
  • Experience working with Software Development Life Cycle (SDLC), Agile methodologies (Scrum)

What we'd like you to have

  • At least 5 years of DataBricks/Sparks Engineering experience
  • Active security clearance 
  • DataBricks Certification or Apache Spark Certification

About BigBear.ai

BigBear.ai is a new leader in decision dominance serving the national defense and intelligence communities. The Company delivers high-end capabilities across the data and digital spectrum to deliver information superiority and decision support. BigBear.ai provides a comprehensive suite of solutions including artificial intelligence and machine learning, data science, advanced analytics, offensive and defensive cyber, data management, cloud solutions, digital engineering, and systems integration. BigBear.ai’s customers, including the U.S. Intelligence Community, Department of Defense, and U.S. Federal Government, rely on its advanced technology solutions to analyze information, manage risk, and solve complex problems, leading to better decision making. Headquartered in Columbia, Maryland, BigBear.ai has additional locations in Virginia, Massachusetts, and California. 


BigBear.ai is an Equal Opportunity/Veterans/Disabled Employer. 

Options

Sorry the Share function is not working properly at this moment. Please refresh the page and try again later.
Share on your newsfeed