Data Engineer Remote Jobs

106 Results

2d

Senior Data Engineer

Expression NetworksDC, US - Remote
agileAbility to travelnosqlsqlDesignscrumjavapythonjavascript

Expression Networks is hiring a Remote Senior Data Engineer

Expression is looking to hire a Senior Data Engineer (individual contributor) to add to the continued growth we are seeing with our Data Science department. This position will daily report to the program manager and data team manager on projects and will be responsible for the design and execution of high-impact data architecture and engineering solutions to customers across a breadth of domains and use cases.

Location:

  • Remote with the ability to travel monthly when needed
    • Local (DC/VA/MD Metropolitan area) is preferred but not required
    • Relocation assistance available for highly qualified candidates

Security Clearance:

  • US Citizenship required
  • Ability to obtain Secret Clearance or higher

Primary Responsibilities:

  • Directly working and leading others on the development, testing, and documentation of software code and data pipelines for data extraction, ingestion, transformation, cleaning, correlation, and analytics
  • Leading end-to-end architectural design and development lifecycle for new data services/products, and making them operate at scale
  • Partnering with Program Managers, Subject Matter Experts, Architects, Engineers, and Data Scientists across the organization where appropriate to understand customer requirements, design prototypes, and optimize existing data services/products
  • Setting the standard for Data Science excellence in the teams you work with across the organization, and mentoring junior members in the Data Science department

Additional Responsibilities:

  • Participating in technical development of white papers and proposals to win new business opportunities
  • Analyzing and providing feedback on product strategy
  • Participating in research, case studies, and prototypes on cutting-edge technologies and how they can be leveraged
  • Working in a consultative fashion to improve communication, collaboration, and alignment amongst teams inside the Data Science department and across the organization
  • Helping recruit, nurture, and retain top data engineering talent

Required Qualifications:

  • 4+ years of experience bringing databases, data integration, and data analytics/ML technologies to production with a PhD/MS in Computer Science/Data Science/Computer Engineering or relevant field, or 6+ years of experience with a Bachelor’s degree
  • Mastery in developing software code in one or more programming languages (Python, JavaScript, Java, Matlab, etc.)
  • Expert knowledge in databases (SQL, NoSQL, Graph, etc.) and data architecture (Data Lake, Lakehouse)
  • Knowledgeable in machine learning/AI methodologies
  • Experience with one or more SQL-on-Hadoop technology (Spark SQL, Hive, Impala, Presto, etc.)
  • Experience in short-release cycles and the full software lifecycle
  • Experience with Agile development methodology (e.g., Scrum)
  • Strong writing and oral communication skills to deliver design documents, technical reports, and presentations to a variety of audiences

Benefits:

  • 401k matching
  • PPO and HDHP medical/dental/vision insurance
  • Education reimbursement
  • Complimentary life insurance
  • Generous PTO and holiday leave
  • Onsite office gym access
  • Commuter Benefits Plan

About Expression:

Founded in 1997 and headquartered in Washington DC, Expression provides data fusion, data analytics, software engineering, information technology, and electromagnetic spectrum management solutions to the U.S. Department of Defense, Department of State, and national security community. Expression’s “Perpetual Innovation” culture focuses on creating immediate and sustainable value for our clients via agile delivery of tailored solutions built through constant engagement with our clients. Expression was ranked #1 on the Washington Technology 2018's Fast 50 list of fastest growing small business Government contractors and a Top 20 Big Data Solutions Provider by CIO Review.

Equal Opportunity Employer/Veterans/Disabled

See more jobs at Expression Networks

Apply for this job

2d

Lead Data Engineer

Logic20/20 Inc.Washington, DC, Remote
agilesqlDesignmongodbazurescrumpythonAWS

Logic20/20 Inc. is hiring a Remote Lead Data Engineer

Job Description

We are looking for a Data Engineer to lead our Data Management team to work on interesting projects to help our clients scale their data solutions to make data driven decisions.   

As a Data Engineer, you’ll work closely with the client to understand both their business processes and analytics needs to design and build data pipelines and cloud data solutions. You will have the opportunity to guide your client through best practices in data lake, data processing, and data pipeline design to help them achieve their business goals. 

You will collaborate with your team including analysts, dashboard developers, and technical project managers to design solutions and work together to deliver a world class solution.   

The ideal candidate will have the balance of technical skills and business acumen to help the client better understand their core needs while understanding technical limitations.   

About you:

  • Collaborative partner who can patiently communicate at the appropriate level to both business and technology teams to understand business needs and pain points 
  • Creative in meeting the client’s core needs with their technology  
  • Determined and able to manage obstacles while maintaining a positive outlook 
  • Adept at explaining technical benefits and deficits to non-technical audiences 
  • Self-driven lifelong learner passionate about learning new data tools and best practices  

What we offer our consultants:   

  • Working on challenging, impactful projects that push the edge of technology 
  • Experience working with both large enterprise clients and mid-sized clients  
  • Progressive responsibilities that encourage ownership and practice leadership  
  • Opportunity to learn and gain experience in complimentary skills such as meeting facilitation, big data processing, project management, data science, and visual analytics  
  • Training and certification opportunities to support your career now and after Logic20/20  
  • Various opportunities to give back to the community through company sponsored events  

Qualifications

Must have:

  • 8+ years of cloud data engineering experience 
  • 5+ years of experience designing and developing cloud and data pipelines with dbt
  • Strong experience with Snowflake
  • Strong experience and knowledge of Python or R
  • Demonstrated knowledge of SQL, creating advanced queries in a high-impact role 
  • Understanding of data modeling (DataVault2.0) 
  • Experience in working with business to understand the appropriate data model (relational, tabular, transactional) for their data solution
  • Deep experience designing and building ELT jobs to move and transform data from various source types and performing exploratory data analysis, data cleansing, and aggregation 
  • Experience scaling & automation to data preparation techniques
  • Experience finetuning ELT to optimize performance 
  • Experience with delivering and supporting production data systems 
  • Experience with developing and operating CI/CD pipelines and other DataOps fundamentals 
  • Experience developing client facing, core design documents: data flows, source to target, requirements, data lineage, and data dictionary
  • Excellent foundation of consulting skills: analytical, written and verbal communication, and presentation skills  
  • Demonstrated ability to identify business and technical impacts of user requirements and incorporate them into the project schedule  
  • Knowledge and exposure to configuring different cloud environments in Azure, AWS, and GCP 
  • You have demonstrated experience in Agile methodologies 
  • Ability to work both independently and as part of a team   
  • Experience working in ambiguous and dynamic environments that move quickly 

Preferred:

  • An undergraduate degree in technology or business  
  • Experience with AutomateDV or datavault4dbt
  • Wherescape DataVault Express (3D and RED)
  • Experience with data warehouse/lakehouse development
  • Microsoft Certifications (Azure Fundamentals, Azure Data Engineer Associate, MCSA [SQL BI Development, Database Development, Data Management + Analytics, BI Reporting], Power Platform) 
  • AWS Certifications (Data Analytics Specialty, Database Specialty, Machine Learning Specialty, Cloud Practioner, Solutions Architect) 
  • Snowflake Certifications (SnowPro Core, SnowPro Advanced Data Engineer) 
  • Experience with Big Data Technologies (Hadoop, Spark, MongoDB) 
  • Experience and certifications with Agile, Scrum, and/or SAFe 

See more jobs at Logic20/20 Inc.

Apply for this job

2d

Senior Data Engineer

Flipside CryptoBoston, MA - Remote
postgressqlpythonAWS

Flipside Crypto is hiring a Remote Senior Data Engineer

Flipside is a commercialization engine that powers blockchain growth. We activate 100,000+ analysts to produce insights, drive narratives, and transact, building emotional connection and a strong and active community that supports token price and chain growth. We make that possible by curating best-in-class data and making that data available entirely for free. Founded in 2017, Flipside works with partners such as Solana, Near, Flow, Avalanche, SEI, Aptos, Blast, Vertex. The company is backed by leading investors including Republic Capital, True Ventures, Galaxy Digital, M13, and more.

Description

As a Senior Data Engineer at Flipside, you will serve in a highly cross functional role and be responsible for creating innovative data platform features that power Flipside’s analytics offerings. You will be part of a high performing team working on a novel data platform using DBT, python, and AWS. This role is critical to the success of Flipside.

Responsibilities

  • Write SQL to query Snowflake to assess operational performance and data quality issues. Familiarity with dbt is a plus.
  • Partner with Analytics to build data pipelines for data ingestion following ELT best practices.
  • Build python data pipelines using AWS lambda and chalice to ingest blockchain data from various APIs. CDK knowledge is a plus.
  • Build CI/CD pipelines using Github Actions.
  • Monitor and maintain existing data pipelines.
  • Build data apis, using python, to integrate into our data platform.
  • Use orchestration tools such as HEVO and FiveTran.
  • Optimizing poor performing Snowflake queries.

Requirements

  • You possess a strong analytical background that includes experience working with complex data sets.
  • You have experience working with and are a subject matter expert with Blockchain Data.
  • You are comfortable working independently with little supervision and thrive in ambiguity.
  • You have a strong background in data modeling and are familiar with star schema and dimensional modeling.
  • You are self directed and capable of driving projects forward to meet deadlines.
  • You have great communication skills (verbal and written) and are able to present your ideas to a wide audience.
  • You have a strong background in writing SQL (Snowflake, Athena, Big Query, or Postgres).

See more jobs at Flipside Crypto

Apply for this job

3d

Data Engineer

ThumbtackRemote, United States
tableauairflowsqlDesignpython

Thumbtack is hiring a Remote Data Engineer

A home is the biggest investment most people make, and yet, it doesn’t come with a manual. That's why we’re building the only app homeowners need to effortlessly manage their homes —  knowing what to do, when to do it, and who to hire. With Thumbtack, millions of people care for what matters most, and pros earn billions of dollars through our platform. And as one of the fastest-growing companies in a $600B+ industry — we must be doing something right. 

We are driven by a common goal and the deep satisfaction that comes from knowing our work supports local economies, helps small businesses grow, and brings homeowners peace of mind. We’re seeking people who continually put our purpose first: advocating for pros and customers, embracing change, and choosing teamwork every day.

At Thumbtack, we're creating a new era of home care. If making an impact and the chance to do good inspires you, join us. Imagine what we’ll build together. 

Thumbtack by the Numbers

  • Available nationwide in every U.S. county
  • 80 million projects started on Thumbtack
  • 10 million 5-star reviews and counting
  • Pros earn billions on our platform
  • 1000+ employees 
  • $3.2 billion valuation (June, 2021) 

About the Engineering Team

At Thumbtack, engineers at every level build products and systems that directly impact our customers and professionals. Our challenges span a wide variety of areas, ranging from building search and booking experiences, to optimizing pricing systems, to building tools to help professionals grow their businesses. We believe in tackling these hard problems together as a team, with strong values around collaboration, ownership and transparency. To read more about the hard problems that our team is taking on, visit our engineering blog.

About the Role

As a Data Engineer, you will work closely with product and engineering teams throughout Thumbtack, and turn data into insight into action. We’re building a hybrid-embedded team of Data Engineers that consult directly with teams to help them better model the data necessary to build a better marketplace for local services. You’ll work to understand requirements, then design, deploy, test, and deploy data pipelines, as well as help architect the core data ecosystem for Analysts, Machine Learning Engineers, and Data Scientists. Major project areas include: designing a next-gen warehouse with DBT, developing advanced ingress/egress validation in the data lake, and modeling cost of supply acquisition for our two-sided marketplace.

Responsibilities

  • Design, build, and maintain data pipelines that support a rapidly evolving product, in partnership with our stellar Analytics and Data Services teams.
  • Build and own the reporting pipelines that organize and structure our terabytes of data into digestible tables that empower analysis and reporting across the company
  • Integrate with product teams throughout the org to understand their data needs, and deliver pipelines and insights with the same engineering rigor as any other software we design.
  • Drive data quality across key product and business areas.
  • Collaborate closely with analysts and engineers to ensure our analytics infrastructure helps meet company goals and allows product development to move quickly.
  • Partner with data platform and product engineering teams to ensure consistent, seamless tracking and measurement of key company and product metrics.

What you’ll need

If you don't think you meet all of the criteria below but still are interested in the job, please apply. Nobody checks every box, and we're looking for someone excited to join the team.

  • 2 or more years of experience working in data engineering or a similar discipline, where your primary focus was on datastores and business intelligence, serving analysis and reporting functions.
  • Experience building ETL data pipelines in a programming language like Python, SQL, or Scala.
  • Expertise in SQL for analytics/reporting/business intelligence, including basic comprehension of query execution plans, and query tuning/optimization for Google BigQuery, Amazon Redshift, Snowflake, or similar.
  • Experience designing, architecting, and maintaining a data warehouse that seamlessly stitches together data from production databases, clickstream event data, and external APIs to serve teams of analysts.
  • Excellent ability to understand the needs of and collaborate with stakeholders in other functions, especially an Analytics team, and identify opportunities for process improvements.

Bonus points if you have 

  • Experience using and/or configuring Business Intelligence tools (Tableau, Mode, Looker, et al).
  • Experience working with semi- or unstructured data in a data lake.
  • Experience working in data engineering or a similar discipline at an online marketplace or similar consumer technology company.
  • Experience orchestrating data pipelines that serve hourly or daily metric reporting in Airflow or similar.

Thumbtack is a virtual-first company, meaning you can live and work from any one of our approved locations across the United States, Canada or the Philippines.* Learn more about our virtual-first working modelhere.

For candidates living in San Francisco / Bay Area, New York City, or Seattle metros, the expected salary range for the role is currently $150,000 - $209,000. Actual offered salaries will vary and will be based on various factors, such as calibrated job level, qualifications, skills, competencies, and proficiency for the role.

For candidates living in all other US locations, the expected salary range for this role is currently $145,000 - $178,000. Actual offered salaries will vary and will be based on various factors, such as calibrated job level, qualifications, skills, competencies, and proficiency for the role.

#LI-Remote

Benefits & Perks
  • Virtual-first working model coupled with in-person events
  • 20 company-wide holidays including a week-long end-of-year company shutdown
  • Library (optional use collaboration & connection hub)in San Francisco
  • WiFi reimbursements 
  • Cell phone reimbursements (North America) 
  • Employee Assistance Program for mental health and well-being 

Learn More About Us

Thumbtack embraces diversity. We are proud to be an equal opportunity workplace and do not discriminate on the basis of sex, race, color, age, pregnancy, sexual orientation, gender identity or expression, religion, national origin, ancestry, citizenship, marital status, military or veteran status, genetic information, disability status, or any other characteristic protected by federal, provincial, state, or local law. We also will consider for employment qualified applicants with arrest and conviction records, consistent with applicable law. 

Thumbtack is committed to working with and providing reasonable accommodation to individuals with disabilities. If you would like to request a reasonable accommodation for a medical condition or disability during any part of the application process, please contact:recruitingops@thumbtack.com

If you are a California resident, please review information regarding your rights under California privacy laws contained in Thumbtack’s Privacy policy available athttps://www.thumbtack.com/privacy/.

See more jobs at Thumbtack

Apply for this job

6d

Senior Data Engineer

Tiger AnalyticsJersey City,New Jersey,United States, Remote

Tiger Analytics is hiring a Remote Senior Data Engineer

Tiger Analytics is a fast-growing advanced analytics consulting firm. Our consultants bring deep expertise in Data Science, Machine Learning and AI. We are the trusted analytics partner for several Fortune 100 companies, enabling them to generate business value from data. Our business value and leadership has been recognized by various market research firms, including Forrester and Gartner. We are looking for top-notch talent as we continue to build the best analytics global consulting team in the world.

We are seeking an experienced Data Engineer to join our data team. As a Data Engineer, you will be responsible for designing, building, and maintaining data pipelines, data integration processes, and data infrastructure using Dataiku. You will collaborate closely with data scientists, analysts, and other stakeholders to ensure efficient data flow and support data-driven decision making across the organization.

  • 8+ years of overall industry experience specifically in data engineering
  • Strong knowledge of data engineering principles, data integration, and data warehousing concepts.
  • Strong understanding of the pharmaceutical/ Life Science domain, including knowledge of patient data, Commercial data, drug development processes, and healthcare data.
  • Proficiency in data engineering technologies and tools, such as SQL, Python, ETL frameworks, data integration platforms, and data warehousing solutions.
  • Experience with data modeling, database design, and data architecture principles.
  • Familiarity with big data technologies (e.g., Hadoop, Spark) and cloud platforms - AWS, Azure
  • Strong analytical and problem-solving skills, with the ability to work with large and complex datasets.
  • Strong communication and collaboration abilities.
  • Attention to detail and a focus on delivering high-quality work.

Significant career development opportunities exist as the company grows. The position offers a unique opportunity to be part of a small, challenging, and entrepreneurial environment, with a high degree of individual responsibility.

See more jobs at Tiger Analytics

Apply for this job

8d

Senior Data Engineer

SendleAustralia (Remote)
tableausqlDesigngit

Sendle is hiring a Remote Senior Data Engineer

Sendle builds shipping that is good for the world. We help small businesses thrive by making parcel delivery simple, reliable, and affordable. We’re a B Corp and the first 100% carbon neutral delivery service in Australia, Canada, and the United States, where we harness major courier networks to create a delivery service that levels the playing field for small businesses.

We envision a world where small businesses can compete on a level playing field with the big guys. Sendle is a fast-growing business with bold ambitions and big dreams. 

In the last few years, we’ve made huge strides towards our goal of becoming the largest SMB eCommerce courier in the world, moving from a single-country operation in Australia to a successful launch and operation in the US and Canada.  We’ve also launched major partnerships with Vestiaire Collective, eBay, Shopify, and Etsy too! 

But most importantly, we’re a bunch of good people doing good work. Wanna join us?

A bit about the role

We are looking for a Senior Data Engineer who is passionate about building scalable data systems that will enable our vision of data democratization to drive value for the business.

As a company, data is at the center of every critical business decision we make. With this role, you will work across many different areas of the business, learning about everything from marketing and sales to courier logistics and network performance. Additionally, there is the opportunity to work directly with stakeholders, with you being a technical thought partner and working collaboratively to design and build solutions to address key business questions. 

What you’ll do

  • Develop, deploy, and maintain data models to support the data needs of various teams across the company
  • Build data models with DBT, utilizing git for source control. Ingest data from different sources (via Fivetran, APIs, etc.) into Snowflake for use by the DBT models
  • Collaborate with the Data Engineering team to brainstorm, scope, and implement process improvements
  • Work with the entire Data and Analytics team to enhance data observability and monitoring
  • Act as a thought partner for stakeholders and peers across the company on ad hoc data requests and identify the best approach and design for our near-term and long-term growth objectives
  • Understand the tradeoffs between technical possibilities and stakeholder needs and strive for balanced solutions
  • Hold self and others accountable to meet commitments and act with a clear sense of ownership
  • Demonstrate persistence in the face of obstacles, resolve them effectively, and involve others as needed
  • Contribute to our data literacy efforts by improving the accessibility, discoverability, and interpretability of our data
  • Research industry trends and introduce new methodologies and processes to the team

What you’ll need

  • Experience with data modeling, data warehousing, and building ETL pipelines (Dagster, DBT, and Snowflake experience a plus)
  • Advanced SQL knowledge
  • Experience with source control technologies such as git
  • Strong communication skills and the ability to partner with business stakeholders to translate business requirements into technical solutions
  • Ability to effectively communicate technical approach with teammates and leaders 
  • Ability to thrive in a remote environment through effective async communication and collaboration
  • Ability to manage multiple projects simultaneously
  • A can-do attitude and demonstrates flexibility by readily taking on new opportunities and assisting others
  • The 5Hs (our core values) in their approach to work and building partnerships with stakeholders and teammates

What we’re offering

  • The chance to work with a creative team in a supportive environment
  • A personal development budget
  • You are able to create your own work environment, connecting to a remote team from anywhere in Australia
  • EAP access for you and your immediate family, because we care about your wellbeing
  • Options through participation in Sendle’s ESOP

What matters to us

We believe that our culture is one of our most important assets. We have 5 key values that we look for in every member of our team.

  • Humble - We put others first. We embrace and seek feedback from others.
  • Honest - We speak gently but frankly. We take ownership of our mistakes and speak the truth.
  • Happy - We enjoy the journey. We are optimistic and find opportunities in all things.
  • Hungry -We aspire to make a difference. We aim high, step out of our comfort zones, and tackle the hard problems.
  • High-Performing - We relentlessly deliver.We know the goal and work fearlessly towards it.

Legally, we need you to know this: 

We are an equal opportunity employer and value diversity. We do not discriminate on the basis of race, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.

If you require accommodations due to a disability to participate in the application or interview process, please get in touch with our team at careers@sendle.com to discuss your needs.

But it’s important to us that you know this:

We strongly believe that diversity of experience contributes to a broader collective perspective that will consistently lead to a better company and better outcomes. We are working hard to increase the diversity of our team wherever we can and we actively encourage everyone to consider becoming a part of it.

If you want to be a part of something remarkable then we’re excited to hear from you.

Interested in knowing more? Check us out on our Careers Page, Being Carbon Neutral and LinkedIn.

 

#LI-Remote

See more jobs at Sendle

Apply for this job

9d

Full Stack Data Engineer

Blue Orange DigitalBogotá, CO - %LABEL_POSITION_TYPE_REMOTE_ANY%
Bachelor's degreesqlapipythonAWS

Blue Orange Digital is hiring a Remote Full Stack Data Engineer

Company Overview:

Blue Orange Digital is a cloud-based data transformation and predictive analytics development firm with offices in NYC and Washington, DC. From startups to Fortune 500s, we help companies make sense of their business challenges by applying modern data analytics techniques, visualizations, and AI/ML. Founded by engineers, we love passionate technologists and data analysts. Our startup DNA means everyone on the team makes a direct contribution to the growth of the company.

Position Overview:

Join our Data Team where we harness the power of data to drive decision-making and support our dynamic research and data services desk. We are looking for a creative and versatile Full Stack Data Engineer to take on exciting projects across various facets of the data stack.

This is a full-time fully remote position for Latin American-based applicants within +/- 2 hours of Eastern Standard Time Zones (NYC). Resume must demonstrate professional English ability.

Responsibilities:

  • Build and prototype scalable data pipelines and build out new API integrations to support continuing increases in data volume and complexity.
  • Write reusable, testable, and efficient code in Python for data ingestion and integration tasks.
  • Use SQL and Python to help find insights into data, collaborating with business stakeholders on specific projects.
  • Use business intelligence tools like Looker and Power BI to create insightful dashboards and visualizations that help stakeholders make data-driven decisions.
  • Work on multiple projects simultaneously, adapting to changing priorities in a fast-paced environment.
  • Act as a problem solver, employing a creative approach to tackle complex challenges and drive continuous improvement in data strategy and execution.

Requirements:

  • Bachelor's degree in Computer Science, Software Engineering, or related field is preferred.
  • Proven experience as a data engineer or in a similar role with a focus on data integration and management.
  • Strong programming skills in Python, with experience in API interactions and automation.
  • Solid foundation in SQL and relational database design.
  • Experience with business intelligence tools such as Looker and Power BI.
  • Excellent problem-solving skills and the ability to think algorithmically.
  • Strong communication skills and the ability to work collaboratively in a team-oriented environment.

Preferred Skills:

  • Experience with cloud services (AWS, Google Cloud Platform, etc.)
  • Knowledge of data warehousing and ETL techniques.
  • Familiarity with version control tools like Git.

Benefits:

  • Fully remote
  • Flexible Schedule
  • Unlimited Paid Time Off (PTO)
  • Paid parental/bereavement leave
  • Worldwide recognized clients to build skills for an excellent resume
  • Top-notch team to learn and grow with

Salary:$5,833.33 to $6,187.50 (monthly salary range)

Background checks may be required for certain positions/projects.

Blue Orange Digital is an equal opportunity employer.

See more jobs at Blue Orange Digital

Apply for this job

FanDuel is hiring a Remote Staff Data Platform Engineer

Job Application for Staff Data Platform Engineer at FanDuel

See more jobs at FanDuel

Apply for this job

10d

Senior Data Engineer

CloudFactoryReading,England,United Kingdom, Remote
tableauterraformsqlpythonAWS

CloudFactory is hiring a Remote Senior Data Engineer

As a global leader in combining people and technology to help companies realise the disruptive potential of AI, CloudFactory is seeking highly skilled and experienced Senior Data Engineers to join our dynamic team to help mature our data platform.

You will have the opportunity to work with cutting-edge technologies and collaborate with a diverse and talented team to scale our data processing capabilities, whilst creating economic opportunities for talented individuals in developing nations.

Although we provide the option to work remotely, we would prefer this position to be office-based for two days per week.

Responsibilities

The successful candidate will translate user requirements into actionable insights by leveraging existing tools and technologies. Your responsibilities will include: 

  • Regular interaction with PMs/stakeholders to understand their requirements
  • Investigation to ascertain if the requirements can be met using existing technology and data resources
  • Data extraction using Python and other tools like Fivetran
  • Data modelling using SQL and the concept of Star and Snowflake schema
  • Assisting in cost and performance optimization choices within the data warehouse.  Communicating reasoning and logic behind opinions to help make optimal choices
  • Write data quality and other metadata collection systems within data pipelines to promote observability and recoverability of pipelines
  • Write unit and integration tests to bolster system reliability
  • Proactively identify the areas of improvement, propose solutions and manage technical debt
  • Mentoring, guiding and line management of three/four team members.

Tech Stack

Our tech stack:

  • Data extraction and orchestration tools - Fivetran, Prefect
  • Data modelling tool - DBT
  • Data Warehouse - Snowflake
  • Data quality cheeks - dbt, Elementary
  • Data visualisation - Tableau and Quicksight
  • Data catalogue and lineage - Atlan, dbt
  • Source code management - Github
  • Infrastructure as Code - Terraform, Cloudformation
  • Programming Language - Python
  • AWS services (ECS, S3)


Interview Process

Candidates will be expected to complete an initial coding and analytical thinking test, followed by an interview with Data Architect and Senior Product Manager where they will be expected to talk through previous projects, including approaches taken and results achieved. 

Successful candidates will then meet with the Engineering Director before a final interview with the CTO.

The process is expected to take 1-2 weeks.

The ideal candidate will possess:

  • Proficiency in requirement capture and technical design
  • Mastery in Data pipeline development
    • Proficiency with extracting data from source systems using Python
    • Mastery in transforming data using SQL and data transformation tools
    • Mastery of data modelling concepts and practices, preferably with Star schema
  • Proficiency in writing data quality tests
  • Proficiency in writing end-to-end tests using test scenarios approach including test data creation and management
  • Proficiency in data pipeline observability techniques (e.g. monitoring, alerting, anomaly detection etc)
  • Some experience in building reports and dashboards
  • Data Warehouse administration
    • Proficiency with Data Warehouse administration (e.g. roles, permissions, compute, performance)
  • Data security and Governance
    • Proficiency in Data security and Governance (e.g. RBAC, encryption, masking, data classification and tagging, time travel and data recovery) practices
    • Some knowledge of using data catalogue and data lineage tools preferably Atlan and dbt documentation
  • Performance and cost optimization
    • Proficiency in performance and cost optimization of database queries using tools and techniques
  • Proficiency with line management to manage a team of Data Engineers and Data Analysts/Analytics Engineers
  • Excellent communication skills, able to translate complex data into understandable insights and narratives.

Location dependent, but can include

  • Office snacks provided Monday-Friday, lunches
  • Home work and allowance
  • Medical
  • Amazing Company Mission and Culture
  • Growth Opportunities


Join us, and change the world for the better. If you are skilled and humble, with a commitment to lifelong learning, and you’re curious about the world and its people, you could be a good fit at CloudFactory. We welcome the unique contributions you can bring to help us build a diverse, inclusive workplace because we connect, learn, and grow stronger from our differences. We want you to bring your whole, authentic self to work.

We look forward to hearing from you!

Still unsure? Read ‘5 Reasons You Should Work at CloudFactory’.

See more jobs at CloudFactory

Apply for this job

10d

Data Engineer (Colombia)

SezzleColombia, Remote
golangBachelor's degreeterraformsqlDesignansiblec++dockerkubernetespythonAWS

Sezzle is hiring a Remote Data Engineer (Colombia)

About the Role: 

We are looking for a Data Engineer who will assist us building, running and improving the data infrastructure that data and engineering teams use to power their services.  Your duties will include the development, testing, and maintenance of data tooling and services, using a combination of cloud products, open source tools and internal applications. You should be able to build high-quality, scalable solutions for a variety of problems. We are seeking a talented and motivated Data Engineer who is best in class with a high IQ plus a high EQ. This role presents an exciting opportunity to thrive in a dynamic, fast-paced environment within a rapidly growing team, with abundant prospects for career advancement.

About Sezzle:

Sezzle is a cutting-edge fintech company dedicated to financially empowering the next generation. With only one in three millennials owning a credit card and the majority lacking their desired credit scores, Sezzle addresses these challenges through a payment platform that offers interest-free installment plans at online stores. By increasing consumers' purchasing power, Sezzle drives sales and basket sizes for thousands of eCommerce merchants that it partners with.

Key Responsibilities Include:

  • Work with a team to plan, design and build tools and services that improve our internal data infrastructure platform and the pipelines that feed it using Python, Go, AWS, Terraform, and Kubernetes.
  • Develop monitoring and alerting of our data infrastructure to detect problems.
  • Perform ongoing maintenance of our data infrastructure, such as apply upgrades.
  • Assist product developers, data scientists and machine learning engineers in debugging and triaging production issues.
  • Collaborate with cross-functional teams to integrate machine learning solutions into production systems
  • Take part in the postmortem reviews, suggesting ways we can improve the reliability of our platform.
  • Document the actions you take and produce both runbooks and automation to reduce day to day toil.

Minimum Requirements:

  • Bachelor's in Computer Science, Data Science, Machine Learning or a related field

Preferred Knowledge and Skills:

  • Experience with AWS services like Redshift, Glue, Sagemaker etc
  • Experience with Data Focused Languages such as Python or SQL
  • Knowledge of ML model training/deploy is a plus
  • Knowledge in data analysis algorithms (e.g. statistics, machine learning)
  • Familiarity with machine learning frameworks and libraries, such as TensorFlow or PyTorch
  • Experience with MLOps principles is a plus
  • Familiarity with orchestration tools like Dagster/Airflow
  • Basic knowledge of Golang, Docker and Kubernetes
  • Familiarity with deployment/provisioning tools like Terraform, Helm, Ansible
  • Experience documenting requirements and specification

About You: 

  • You have relentlessly high standards - many people may think your standards are unreasonably high. You are continually raising the bar and driving those around you to deliver great results. You make sure that defects do not get sent down the line and that problems are fixed so they stay fixed.
  • You’re not bound by convention - your success—and much of the fun—lies in developing new ways to do things
  • You need action - speed matters in business. Many decisions and actions are reversible and do not need extensive study. We value calculated risk-taking.
  • You earn trust - you listen attentively, speak candidly, and treat others respectfully.
  • You have backbone; disagree, then commit - you can respectfully challenge decisions when you disagree, even when doing so is uncomfortable or exhausting. You have conviction and are tenacious. You do not compromise for the sake of social cohesion. Once a decision is determined, you commit wholly.

What Makes Working at Sezzle Awesome? 

At Sezzle, we are more than just brilliant engineers, passionate data enthusiasts, out-of-the-box thinkers, and determined innovators. We believe in surrounding ourselves with only the best and the brightest individuals. Our culture is not defined by a certain set of perks designed to give the illusion of the traditional startup culture, but rather, it is the visible example living in every employee that we hire. Sezzle recognizes and values the importance of diversity and inclusion in enriching the employment experience of its employees and in supporting our mission.

#Li-remote

See more jobs at Sezzle

Apply for this job

11d

Senior Data Engineer

NationsBenefitsDallas, TX Remote
Bachelor's degreeDesignazurepythonAWS

NationsBenefits is hiring a Remote Senior Data Engineer

NationsBenefits is recognized as one of the fastest-growing companies in America and a Healthcare Fintech provider of supplemental benefits, flex cards, and member engagement solutions. We partner with managed care organizations to provide innovative healthcare solutions that drive growth, improve outcomes, reduce costs, and bring value to their members.

Through our comprehensive suite of innovative supplemental benefits, fintech payment platforms, and member engagement solutions, we help health plans deliver high-quality benefits to their members that address the social determinants of health and improve member health outcomes and satisfaction.

Our compliance-focused infrastructure, proprietary technology systems, and premier service delivery model allow our health plan partners to deliver high-quality, value-based care to millions of members.

We offer a fulfilling work environment that attracts top talent and encourages all associates to contribute to delivering premier service to internal and external customers alike. Our goal is to transform the healthcare industry for the better! We provide career advancement opportunities from within the organization across multiple locations in the US, South America, and India.

Role:

As a Senior Data Engineer, you will develop and implement robust data engineering solutions to support our organization's data-driven initiatives. The ideal candidate will have at least five years of hands-on experience in data engineering on high-performing teams. Expertise in DataBricks, dbt, and Python is a critical requirement for this role.

Key Responsibilities:

  • Design, develop, and maintain scalable data pipelines and ETL processes using Databricks, dbt, and Python.
  • Collaborate with cross-functional teams, including data scientists, analysts, and stakeholders, to understand data requirements and deliver solutions that meet business needs.
  • Optimize and tune existing data pipelines for performance, reliability, and scalability.
  • Implement best practices for data governance, security, and compliance.
  • Lead efforts to streamline and automate data processes, reducing manual intervention and increasing efficiency.
  • Mentor and guide junior members of the data engineering team.
  • Stay up-to-date with the latest trends and advancements in data engineering and contribute to continuously improving data engineering practices within the organization.

Qualifications:

  • Bachelor's degree in Computer Science, Information Technology, or a related field.
  • Minimum 5+ years of hands-on experience in data engineering
  • Expertise in building data pipelines and ETL processes using DataBricks, dbt, and Python
  • Extensive experience with PurView a plus.
  • Strong understanding of data warehousing concepts and methodologies.
  • Experience with cloud platforms such as AWS, Azure, or GCP.
  • Excellent communication and interpersonal skills
  • Excellent problem-solving skills and attention to detail.
  • Strong communication and teamwork abilities.
  • Knowledge of data security and compliance standards is a plus.

NationsBenefits is an Equal Opportunity Employer

See more jobs at NationsBenefits

Apply for this job

11d

Director, Data Engineering

NationsBenefitsDallas, TX Remote
Bachelor's degreesqlpython

NationsBenefits is hiring a Remote Director, Data Engineering

NationsBenefits is recognized as one of the fastest-growing companies in America and a Healthcare Fintech provider of supplemental benefits, flex cards, and member engagement solutions. We partner with managed care organizations to provide innovative healthcare solutions that drive growth, improve outcomes, reduce costs, and bring value to their members.

Through our comprehensive suite of innovative supplemental benefits, fintech payment platforms, and member engagement solutions, we help health plans deliver high-quality benefits to their members that address the social determinants of health and improve member health outcomes and satisfaction.

Our compliance-focused infrastructure, proprietary technology systems, and premier service delivery model allow our health plan partners to deliver high-quality, value-based care to millions of members.

We offer a fulfilling work environment that attracts top talent and encourages all associates to contribute to delivering premier service to internal and external customers alike. Our goal is to transform the healthcare industry for the better! We provide career advancement opportunities from within the organization across multiple locations in the US, South America, and India.

Role:

As the Director of Data Engineering, you will oversee the development and implementation of robust data engineering solutions to support our organization's data-driven initiatives. The ideal candidate will possess at least 10 years of hands-on experience in data engineering, with at least 3 years of proven leadership managing high-performing teams. Expertise in DataBricks, dbt, and Python is a critical requirement for this role.

Key Responsibilities:

  • Leadership and Team Management: Lead and inspire a team of data engineers to deliver high-quality, scalable, and efficient data solutions. Provide mentorship, guidance, and professional development opportunities to team members. Foster a collaborative and innovative team culture.
  • Data Engineering Strategy: Develop and execute a comprehensive data engineering strategy through DataBricks Medallion Architecture. Drive innovation in data engineering practices and technologies.
  • Collaboration: Collaborate with cross-functional teams, including data science, analytics, and IT, to ensure seamless integration of data solutions. Work closely with business stakeholders to understand requirements and deliver data solutions that meet their needs.
  • Performance Optimization: Partner with DevOps to identify and address performance bottlenecks in data processing pipelines, ensuring optimal performance and reliability.
  • Quality Assurance: Establish and maintain rigorous testing and quality assurance processes to ensure the accuracy and reliability of data engineering solutions.
  • Documentation: Create and maintain comprehensive documentation for data engineering processes, solutions, and best practices.

Qualifications:

  • Bachelor's degree in Computer Science, Information Technology, or a related field.
  • 10+ years of hands-on experience in data engineering, plus 3 years of team leadership
  • Expert-level knowledge of SQL (ideally DataBricks), dbt, and Python is required.
  • Extensive experience with PurView is a plus.
  • Excellent communication and interpersonal skills
  • Excellent problem-solving skills and attention to detail.
  • Strong communication and teamwork abilities.
  • Knowledge of data security and compliance standards is a plus.

NationsBenefits is an Equal Opportunity Employer

See more jobs at NationsBenefits

Apply for this job

13d

Data Engineer (m/w/d)

GerresheimerEssen, Germany, Remote
azure

Gerresheimer is hiring a Remote Data Engineer (m/w/d)

Stellenbeschreibung

  • Du hilfst uns, den Data Analytics / Business Intelligence Bereich für die Business Unit Moulded Glass aufzubauen und weiterzuentwickeln, um Mehrwert und Business Insights zu generieren.
  • Du entwickelst, implementierst und wartest ETL-Prozesse mit Azure Data Factory und anderen Tools in enger Abstimmung mit unserem Gerresheimer Data Science Center.
  • Du überführst bestehende Auswertungen in standardisierte und automatisierte Lösungen mit Fokus auf Datenmodellierung, und optimierst sie kontinuierlich.
  • Du entwickelst und optimierst Datenmodelle, um einen effizienten Datenzugriff und -analyse zu ermöglichen und integrierst verschiedene Datenquellen und Datenbanken (wie S4 HANA, interne und externe Quellen), um eine konsistente Datenbasis sicherzustellen.
  • Du entwirfst, implementierst und überwachst Datenpipelines, -architektur und Mechanismen zur Datenqualitätsüberwachung und -verbesserung, inklusive Fehlererkennung und -behebung.
  • Du setzt AI/ML-Projekte in enger Zusammenarbeit mit unserem Gerresheimer Data Science Center um.

Die Tätigkeit kann remoteausgeführt werden, falls gewünscht.

Qualifikationen

  • Du verfügst über einen Bachelor- oder Master-Abschluss in Informatik, Mathematik, Ingenieurswissenschaften oder einer verwandten Fachrichtung.
  • Du besitzt fundierte technische Kenntnisse und idealerweise Erfahrungen in der Arbeit mit relevanten Azure-Diensten wie Azure Data Factory, Microsoft Azure, SQL/DWH/Analysis Services, Azure Data Lake, Azure Synapse und Azure DevOps.
  • Erfahrung im Umgang mit Massendaten, Datenmodellierung und Datenverarbeitung sind für diese Position unerlässlich.
  • Du beherrschst mindestens eine Programmiersprache, idealerweise Python.
  • Idealerweise verfügst du über Kenntnisse der Datenstrukturen in ERP-Systemen wie SAP FI/CO/MM/SD.
  • Du hast idealerweise Erfahrungen mit BI-Tools wie Power BI und mit agilen Projektmanagementmethoden wie Scrum.
  • Du bist begeistert von Themen wie Data Analytics, Data Science, AI, Machine und Deep Learning.
  • Du zeigst eine kontinuierliche Lernbereitschaft und Interesse daran, dich auch in anderen Bereichen weiterzubilden (z.B. RPA).
  • Du sprichst Deutsch und Englisch.

See more jobs at Gerresheimer

Apply for this job

14d

Senior Data Engineer

4 years of experienceDesigncssjavascript

WonderBotz is hiring a Remote Senior Data Engineer

Senior Data Engineer - WonderBotz - Career Page

See more jobs at WonderBotz

Apply for this job

15d

Data Engineer (APAC)

In All Media IncPhilippines, PH Remote
agilesqlsalesforcepythonAWS

In All Media Inc is hiring a Remote Data Engineer (APAC)

Data Engineer

In All Media

InallMedia.com is a Global community in charge of allocating and administrating complete teams according to our clients’ needs, always using an agile methodology.

At this moment, we are looking for a Data Engineer. This position is 100% remote and payable in USD.


Role Description

Our client is one of the biggest Job Boards of the world with a presence in 62 countries.

We are looking for Data Engineer with a solid experience in Amazon Web Services.

Must have requirements

  • AWS
  • AWS Glue or EMR
  • SQL
  • Athena
  • Python
  • Spark
  • Redshift
  • S3
  • GitHub

Nice to have requirements

  • Salesforce Service Cloud

Benefits

  • ???? USD Payment
  • ????100% remote
  • ???? Great Community
  • ???? Full-time, long-term
  • ????????Growth opportunities

See more jobs at In All Media Inc

Apply for this job

16d

Senior Data Engineer, Finance

InstacartUnited States - Remote
airflowsqlDesign

Instacart is hiring a Remote Senior Data Engineer, Finance

We're transforming the grocery industry

At Instacart, we invite the world to share love through food because we believe everyone should have access to the food they love and more time to enjoy it together. Where others see a simple need for grocery delivery, we see exciting complexity and endless opportunity to serve the varied needs of our community. We work to deliver an essential service that customers rely on to get their groceries and household goods, while also offering safe and flexible earnings opportunities to Instacart Personal Shoppers.

Instacart has become a lifeline for millions of people, and we’re building the team to help push our shopping cart forward. If you’re ready to do the best work of your life, come join our table.

Instacart is a Flex First team

There’s no one-size fits all approach to how we do our best work. Our employees have the flexibility to choose where they do their best work—whether it’s from home, an office, or your favorite coffee shop—while staying connected and building community through regular in-person events. Learn more about our flexible approach to where we work.

Overview

 

At Instacart, our mission is to create a world where everyone has access to the food they love and more time to enjoy it together. Millions of customers every year use Instacart to buy their groceries online, and the Data Engineering team is building the critical data pipelines that underpin all of the myriad of ways that data is used across Instacart to support our customers and partners.

About the Role 

 

The Finance data engineering team plays a critical role in defining how financial data is modeled and standardized for uniform, reliable, timely and accurate reporting. This is a high impact, high visibility role owning critical data integration pipelines and models across all of Instacart’s products. This role is an exciting opportunity to join a key team shaping the post-IPO financial data vision and roadmap for the company.

 

About the Team 

 

Finance data engineering is part of the Infrastructure Engineering pillar, working closely with accounting, billing & revenue teams to support the monthly/quarterly book close, retailer invoicing and internal/external financial reporting. Our team collaborates closely with product teams to capture critical data needed for financial use cases.

 

About the Job 

  • You will be part of a team with a large amount of ownership and autonomy.
  • Large scope for company-level impact working on financial data.
  • You will work closely with engineers and both internal and external stakeholders, owning a large part of the process from problem understanding to shipping the solution.
  • You will ship high quality, scalable and robust solutions with a sense of urgency.
  • You will have the freedom to suggest and drive organization-wide initiatives.

 

About You

Minimum Qualifications

  • 6+ years of working experience in a Data/Software Engineering role, with a focus on building data pipelines.
  • Expert with SQL and  knowledge of Python.
  • Experience building high quality ETL/ELT pipelines.
  • Past experience with data immutability, auditability, slowly changing dimensions or similar concepts.
  • Experience building data pipelines for accounting/billing purposes.
  • Experience with cloud-based data technologies such as Snowflake, Databricks, Trino/Presto, or similar.
  • Adept at fluently communicating with many cross-functional stakeholders to drive requirements and design shared datasets.
  • A strong sense of ownership, and an ability to balance a sense of urgency with shipping high quality and pragmatic solutions.
  • Experience working with a large codebase on a cross functional team.

 

Preferred Qualifications

  • Bachelor’s degree in Computer Science, computer engineering, electrical engineering OR equivalent work experience.
  • Experience with Snowflake, dbt (data build tool) and Airflow
  • Experience with data quality monitoring/observability, either using custom frameworks or tools like Great Expectations, Monte Carlo etc

 

#LI-Remote

Instacart provides highly market-competitive compensation and benefits in each location where our employees work. This role is remote and the base pay range for a successful candidate is dependent on their permanent work location. Please review our Flex First remote work policyhere.

Offers may vary based on many factors, such as candidate experience and skills required for the role.Additionally, this role is eligible for a new hire equity grant as well as annual refresh grants. Please read more about our benefits offeringshere.

For US based candidates, the base pay ranges for a successful candidate are listed below.

CA, NY, CT, NJ
$192,000$213,000 USD
WA
$184,000$204,000 USD
OR, DE, ME, MA, MD, NH, RI, VT, DC, PA, VA, CO, TX, IL, HI
$176,000$196,000 USD
All other states
$159,000$177,000 USD

See more jobs at Instacart

Apply for this job

16d

Marketing Data Engineer

SamsaraRemote - US
sqlpython

Samsara is hiring a Remote Marketing Data Engineer

Job Application for Marketing Data Engineer at Samsara

Apply for this job

16d

Data Analyst/Engineer

Razorhorse CapitalIndia Remote
Master’s DegreesqlB2BDesignapi

Razorhorse Capital is hiring a Remote Data Analyst/Engineer

Company Overview

Founded in 2012, Razorhorse Capital is a buy-side advisory firm in Austin with remote workers across the world. Razorhorse sources and executes B2B SaaS investments & acquisitions for 20+ Partners in North America, the UK, Europe, and Australia. We are led by senior dealmakers with deep track records in software buyout and growth equity transactions. Our Partners deploy $3B+ capital annually, and our experience closing 100+ deals across a variety of markets, structures, and strategies demonstrates our flexibility and prowess structures.

About the role

Razorhorse is seeking a Data Analyst/Engineer to play a pivotal role in enhancing our data-driven decision-making processes. In this key position, you will develop and maintain advanced ELT processes, design impactful data models and dashboards, and leverage machine learning to drive strategic insights. Your expertise will ensure the integrity and enhancement of our data infrastructure, supporting cross-functional teams in a dynamic, innovative environment. This role is ideal for those passionate about using their technical skills to influence real-world business outcomes.

Key Responsibilities

  • Data Infrastructure: Enhance decision-making capabilities by optimizing data infrastructure.
  • ELT Processes: Develop and maintain advanced ELT processes within Google BigQuery and our ELT tools.
  • Data Modeling: Construct and manage efficient data models for streamlined analysis and reporting.
  • Dashboard Development: Design and implement visually compelling dashboards using Google Looker Studio and Looker platform.
  • Cross-functional Collaboration: Work alongside various teams to define KPIs and craft analytical frameworks.
  • API Integration: Integrate and manage data APIs for seamless connectivity with third-party systems to cover our data gaps and completeness
  • Data Validation: Ensure the accuracy and reliability of data through rigorous validation processes.
  • CRM Data Management: Execute data cleanup and management tasks within our CRM systems.
  • Machine Learning: Utilize and enhance ML models, leveraging Google ML and BigQuery ML for sophisticated analytics.

Qualifications and Skills

  • Educational Background - Bachelor’s or Master’s degree in Computer Science, Statistics, or a related field.
  • Professional Experience - Demonstrable experience as a Data Analyst/Engineer, preferably in a fast-paced, data-intensive environment.
  • Technical Proficiency
    • Expertise in SQL, data modeling, and database design.
    • Proficient with Google Cloud Platforms, including BigQuery, Data Studio, and Looker.
    • Skilled in developing data API integrations.
    • Knowledgeable in programming languages such as Python/R or Java.
    • Familiarity with machine learning concepts and practical experience with ML models.
  • Core Competencies
    • Strong analytical and problem-solving skills.
    • Excellent communication and teamwork capabilities.

Application Process

Location: Global, Remote

Start: Immediate

Position Type: Full-time

Compensation: between base and bonus, OTE is USD 40-50K+

Please apply with your resume and cover letter

Join Razorhorse Capital and play a key role in shaping the future of investment strategies through expert data analysis and engineering.

See more jobs at Razorhorse Capital

Apply for this job

17d

Data Engineer

kea.Brazil Remote
4 years of experienceBachelor's degreepythonAWSjavascript

kea. is hiring a Remote Data Engineer

About kea

Launched in 2018, kea is one of the fastest-growing companies changing the way restaurants operate. We've raised $29 million from Tier 1 Silicon Valley investors, and we're not stopping there. Originating in Mountain View, California, and now fully remote, we are building a team that is excited about our mission: drive the world’s commerce by empowering restaurants to operate more intelligently and own their relationship with customers. The first product in our platform, kea Voice, takes orders over the phone at a high level of efficiency and consistency so that restaurants can focus on making delicious food. That's a win-win!

Our Commitment as a team

At kea, we live our values--we are committed to innovation while moving as one unit at high velocity. We are compassionate and care for others. We are open and transparent. We play to win the long game. As you review the below keep Our Commitment in mind. Your uniqueness is valued and differences of opinion, background, and ability are welcomed. If you think you satisfy even just a few of the below points, please apply!

About the role

You will be joining the autosquad team, focusing on automating the order-taking process through AI. Your role will be pivotal in storing, organizing, and transforming data to train and experiment with new models, ensuring our AI can handle increasingly complex orders.

What your day-to-day looks like

  • Designing, constructing, and managing data warehouses and lakes
  • Ensuring data quality and accessibility by cleaning and preprocessing datasets
  • Developing and maintaining ETL pipelines in a cloud environment
  • Collaborating closely with ML Ops, ML, and software engineers to support data access for model experimentation

What sets you up for success

  • Basic experience in data storage, data manipulation, and programming
  • At least 4 years of experience as a data engineer
  • A Bachelor's degree in Computer Science, Data Science, Physics, Mathematics, or a related field
  • Proficiency in data modeling, warehousing, SQL/NoSQL databases, Apache Spark or Flink, Pandas, NumPy, cloud services (AWS, GCP), and DevOps practices
  • Software engineering skills in Python, JavaScript, and familiarity with CI/CD tools

What we offer

  • Generous stock option plan
  • 100% employer-paid medical and mental health care benefits. Vision and dental are also available
  • Flexible time off
  • Remote work
  • 401K
  • Ownership, responsibility, and empowerment in what you do
  • Incredible teammates and a caring workplace culture

See more jobs at kea.

Apply for this job

18d

Director, Data Engineering

VeriskJersey City, NJ, Remote
5 years of experiencesqlpythonAWS

Verisk is hiring a Remote Director, Data Engineering

Job Description

As the Data Engineering Director, you will lead a team of data engineers responsible for building efficient, maintainable production-grade data pipelines. This role can be based in our Jersey City, NJ location (hybrid work model) or remote.

  • This is a hands-on role with a small team, where you will have the opportunity to provide technical leadership and serve as the bridge between our Data and Product Engineering organizations
  • Lead and grow a small team of Data Engineers, including career and performance management
  • Partner with the Analytics Engineering team build and maintain the roadmap for our data stack and infrastructure
  • Identify existing data capability and infrastructure gaps or opportunities within and across initiatives and provide subject matter expertise in support of remediation 
  • Be a champion of data privacy and quality

Qualifications

  • Bachelor’s Degree in Computer Science, Information Systems, or related field required. Advance Degree preferred
  • 10+ years of experience as data engineer, or other data-related role; at least 4 years of management experience
  • Minimum 2 years experience in the entity resolution space
  • Strong Python and SQL skills and knowledge of data modeling, ETL/ELT development principles, data warehousing concepts, and software engineering concepts
  • Minimum 5 years of experience with Amazon Web Services (AWS) – EC2, ECS, S3, SQS, RDS, Dynamo DB, Cloud Watch, IAM and Lambda.
  • Expertise with Gitlab and CI/CD pipelines
  • Experience managing a cloud data warehouse and a workload scheduling and orchestration framework
  • Strong ability to understand and document critical data through effective data collection, data analysis and data interpretation. 
  • A demonstrated self-starter with strong communication and project management skills

#LI-LM03
#LI-Hybrid

See more jobs at Verisk

Apply for this job