Capstone Integrated Solutions Logo

Capstone Integrated Solutions

Senior Data Engineer (AWS)

Posted 17 Days Ago
Remote
Hiring Remotely in USA
Senior level
Remote
Hiring Remotely in USA
Senior level
Lead data architecture, develop ETL pipelines using AWS Glue, integrate with CMIC ERP, and ensure data quality for AI-driven projects.
The summary above was generated by AI

Capnexus is a comprehensive services provider. Our team consists of outstanding professionals, highly experienced in designing, building, and supporting retail software. We see ourselves as a build-as-a-service provider who follows a repeatable business pattern that can be applied to a variety of platforms and verticals. Having a culture built on outcomes and delivery at the core of the business, Capnexus is providing its customers with a complete suite of services for software development, system analysis, integration, implementation, and support, as well as the option to engage a single team to perform all the services they require. 

Who You Are and What You'll Do: 

Capnexus is looking for a highly skilled Senior AWS Data Engineer to lead data architecture, pipeline development, and ERP integration for a 12-week AI-powered modernization engagement in the construction industry. This role is focused on designing and implementing the data engineering backbone of an intelligent subcontractor pre-qualification platform, including CMIC ERP API integration, Amazon Textract data extraction pipelines, ETL development using AWS Glue, and data quality validation. This is an exciting opportunity to apply advanced cloud data engineering skills on a platform that leverages generative AI to automate and modernize enterprise workflows. 

Responsibilities: 

  • Lead data architecture design, API assessment, and ETL requirements gathering during the Discovery & Design phase. 
  • Develop and configure CMIC ERP API integration to establish reliable data exchange between the ERP system and the AWS platform. 
  • Design and implement data pipelines using AWS Glue for ETL processing of subcontractor documents and ERP data. 
  • Integrate Amazon Textract to extract structured data from insurance certificates, bonding letters, and financial documents. 
  • Build and maintain data models to support AI-powered validation, risk profiling, and executive reporting. 
  • Configure Amazon S3 data lake architecture to store and manage raw, processed, and curated data assets. 
  • Implement AWS Lambda and AWS Step Functions to orchestrate data workflows and automated processing pipelines. 
  • Develop and expose data through Amazon API Gateway to support application and dashboard consumption. 
  • Ensure data quality, validation, and integrity across all integration points and pipeline outputs. 
  • Conduct data integration testing and support user acceptance testing (UAT) for data-dependent features. 
  • Collaborate with Full Stack, AI/ML, and DevOps team members to ensure seamless end-to-end data flows. 
  • Contribute to knowledge transfer documentation, data pipeline runbooks, and operations guides. 

 

Qualifications: 

  • 5+ years of data engineering experience, with at least 2+ years working in AWS cloud environments. 
  • Strong proficiency in SQL and experience with AWS database services including RDS, DynamoDB, and Aurora. 
  • Hands-on experience with AWS Glue for ETL development and data pipeline orchestration. 
  • Experience integrating with ERP systems or enterprise APIs, with ability to assess and document API capabilities. 
  • Working knowledge of Amazon Textract for automated document data extraction. 
  • Proficiency with Amazon S3 for data lake design, storage management, and lifecycle policies. 
  • Experience with AWS Lambda and AWS Step Functions for serverless data workflow orchestration. 
  • Solid understanding of data modeling, data quality validation, and integration testing practices. 
  • Familiarity with Amazon API Gateway for exposing data services to downstream consumers. 
  • Strong analytical, problem-solving, and communication skills with the ability to work in Agile/Scrum teams. 

 

Nice to Have: 

  • Experience with Amazon Bedrock or AI/ML-integrated data pipelines. 
  • Familiarity with CMIC ERP or similar construction industry enterprise resource planning systems. 
  • Experience with GraphQL APIs or AWS AppSync. 
  • Knowledge of TypeScript or Python for pipeline scripting and Lambda function development. 
  • AWS Certification (Data Analytics Specialty, Database Specialty, or Solutions Architect). 
  • Background in construction, insurance, or financial services industries. 

 

"Our Culture": 

At Capstone, the central principles that we all adhere to, and the glue that holds us together, are our keystones. Our four keystones are: 

"A Customer Obsessed, Delivery Focused, Culture" 

  • We’re driven to exceed our customers’ expectations by listening, leading, solving problems, and delivering what we promise 
  • We aim to be the most dependable and trusted partner serving our customers. TRUST = CONSISTENCY x TIME 

"A Culture of Learning and Sharing" 

  • We value “Lifetime Learners”; those who are hungry, competitive, curious, and self-motivated in their pursuit of knowledge. 
  • Personal and professional growth depends on teamwork and continuous learning. By sharing knowledge, skills, ideas, and effort, we benefit our customers, ourselves, and our communities. 
  • We recognize that the thoughts, feelings, and backgrounds of others are as important as our own. Everyone has something to learn and everyone has something they can teach. 
  • Knowledge and ability are valued. Sharing knowledge and helping others learn new capabilities is valued exponentially. 

"A Culture of Growth and Scalability" 

  • Growth comes from not establishing barriers in your role. “Cross functional skill sets are valued and help us deliver to our customers in a truly agile fashion. It comes with understanding that when asked to do something new, you will need support, have questions, and make some mistakes along the way. 
  • The most elegant solution is a simple solution. Simple doesn’t mean easy. It’s often more difficult to break a complex problem down into simple, scalable terms. We don’t appreciate, or value, over architected solutions or superfluous coding. 
  • Time is one of our most precious commodities. Scalability implies being respectful of this and passionate about making the most efficient use of each and every one of our team members time. 

"All Work is Strategic" 

  • No matter how small a project or assignment appears, every single engagement is an opportunity for us to prove ourselves, build trust, and develop relationships that last and grow 
  • Every task, interaction, and commitment matters 
  • Big or small, we execute our plans and strategies with focus, commitment, and passion 

 

We offer: 

Job Type: Full-time, 1099

Duration: Short-term

Benefits: 

  • Remote work 

 

Capnexus is an equal opportunity employer. We embrace and celebrate diversity and are committed to creating an inclusive and safe environment for all employees. Experience comes in many forms, and we’re dedicated to adding new perspectives to the team. We encourage you to apply even if your experience doesn’t perfectly align with what we have listed. We look forward to hearing from you. 

No Agencies Please! 

Similar Jobs

12 Days Ago
In-Office or Remote
Senior level
Senior level
Information Technology • Software
The Senior Data Platform Engineer leads the design and management of data pipelines using Python and AWS, focusing on data engineering for finance-related analytics and systems.
Top Skills: AirflowAWSCdkCloudFormationDockerEcsEventbridgeFastapiFlaskGlueKinesisKubernetesLambdaNumpyPandasPrefectPysparkPytestPythonS3SqlalchemyStep FunctionsTerraform
Yesterday
Remote
USA
Senior level
Senior level
Artificial Intelligence • Information Technology • Software • Database
Build production-grade AI-powered data tooling: extract data from Snowflake, generate and store embeddings, enable semantic search, design enrichment pipelines using LLM APIs, optimize AWS infrastructure, and create reusable services and SDKs for scalable, observable data and AI workflows.
Top Skills: AWSEcsEksEmbeddingsIamLambdaLlm ApisOpenaiPineconePythonS3Semantic SearchSnowflakeVector Databases
6 Days Ago
In-Office or Remote
Mid level
Mid level
Database
As an AWS Data Engineer, you will develop data pipelines, create data models, troubleshoot issues, and collaborate with teams while utilizing tools like PySpark and SQL.
Top Skills: AirflowAmazon AthenaAmazon EmrAmazon GlueAmazon RedshiftPrestoPysparkSnowflakeSQL

What you need to know about the Los Angeles Tech Scene

Los Angeles is a global leader in entertainment, so it’s no surprise that many of the biggest players in streaming, digital media and game development call the city home. But the city boasts plenty of non-entertainment innovation as well, with tech companies spanning verticals like AI, fintech, e-commerce and biotech. With major universities like Caltech, UCLA, USC and the nearby UC Irvine, the city has a steady supply of top-flight tech and engineering talent — not counting the graduates flocking to Los Angeles from across the world to enjoy its beaches, culture and year-round temperate climate.

Key Facts About Los Angeles Tech

  • Number of Tech Workers: 375,800; 5.5% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Snap, Netflix, SpaceX, Disney, Google
  • Key Industries: Artificial intelligence, adtech, media, software, game development
  • Funding Landscape: $11.6 billion in venture capital funding in 2024 (Pitchbook)
  • Notable Investors: Strong Ventures, Fifth Wall, Upfront Ventures, Mucker Capital, Kittyhawk Ventures
  • Research Centers and Universities: California Institute of Technology, UCLA, University of Southern California, UC Irvine, Pepperdine, California Institute for Immunology and Immunotherapy, Center for Quantum Science and Engineering

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account