Data Engineer

Sorry, this job was removed at 10:38 a.m. (PST) on Thursday, June 27, 2019
Find out who's hiring in Santa Monica.
See all Data + Analytics jobs in Santa Monica
Apply
By clicking Apply Now you agree to share your profile information with the hiring company.

At Tradesy, we care deeply about the environmental issues caused by low quality disposable products that just end up polluting our environment. We are a peer-to-peer marketplace that is pioneering a sustainable future for commerce in the fashion industry. We are financially backed by top tier venture capital firms, with millions of passionate members and a product that people love. Our mission is to make the resale of consumer goods as simple, safe, and stylish as retail, at scale and we do this in an  office with an ocean view in sunny Santa Monica, California!

 

Tradesy is seeking an exceptionally talented Data Engineer who works closely with other engineers, data scientists, architects and product owners to develop reliable, efficient, and scalable systems and data pipelines with a strong focus on engineering quality. The ideal candidate has a passion for building things “right”, creating leverage for the business by designing and implementing systems that are low maintenance and withstand the test of time.

 

We are also migrating our applications, data warehouse and data pipelines to Google Cloud to take advantage of Google’s cutting edge technologies and managed services. We’re already leveraging the power of PubSub, BigQuery and Google’s Data ecosystem in very interesting ways. If this sounds exciting to you and you believe you can impact our business with your passion and skills, then you are the right candidate for us.

You Will:

  • Interact and integrate with internal and external teams and systems to extract, transform, and load data from a wide variety of sources
  • Support our BI, Data Science and Engineering teams to assist with data-related technical issues and support their data and data infrastructure needs
  • Migrate data pipelines from AWS to GCP
  • Respond to alerts and investigate data issues
  • Participate in architectural decisions

You Have:

  • Computer Science degree or equivalent experience.
  • 2+ years of experience with data pipeline development (ETL) and data cleansing with large-scale, complex data sets
  • 2+ years of experience with batch processing technologies (ie. Hadoop, Spark, Apache Beam, Flink)
  • Coding proficiency in at least one of Java, Python, C++, Go, Scala
  • Strong computer science skills with a focus on algorithms and data structures
  • A master of data querying for relational and/or NoSQL data stores
  • Experience with multiple data formats: CSV, JSON, Avro, Parquet

Bonus Points:

  • Experience with stream processing technologies such as Apache Beam, Spark Streaming, Flink, Kafka Streams
  • ETL experience on AWS using EMR, Firehose, Lambda
  • ETL experience on Google Cloud using Composer, Dataproc, Cloud Functions, Dataflow
  • Redshift or BigQuery experience
  • Familiar with messaging systems such as Kinesis, Kafka, PubSub
  • Familiar with automation tools such as Apache Airflow, Luigi, AWS Data Pipeline

Compensation:

  • Competitive base salary plus meaningful equity
  • Comprehensive benefits (Medical, Dental, Vision, 401k)
  • Flexible Paid Time Off

Additional Perks:

  • Daily catered lunches
  • Dog friendly office
  • Collaborative, fun team
Read Full Job Description
Apply Now
By clicking Apply Now you agree to share your profile information with the hiring company.

Location

Tradesy is a remote first company with employees based all across the United States.

Similar Jobs

Apply Now
By clicking Apply Now you agree to share your profile information with the hiring company.
Learn more about TradesyFind similar jobs