0% found this document useful (0 votes)
26 views

Data EngineerTO

This job posting seeks a data engineer to build and optimize data pipelines and infrastructure to support products and initiatives. Responsibilities include pipeline building, optimization, and operationalization as well as designing self-serving data infrastructure. Qualifications include experience with Python, SQL, AWS, and building big data pipelines.

Uploaded by

igor.cleto
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
26 views

Data EngineerTO

This job posting seeks a data engineer to build and optimize data pipelines and infrastructure to support products and initiatives. Responsibilities include pipeline building, optimization, and operationalization as well as designing self-serving data infrastructure. Qualifications include experience with Python, SQL, AWS, and building big data pipelines.

Uploaded by

igor.cleto
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

Title: Data Engineer

Location: Remote

Overview

Seeking a creative and savvy Data Engineer to join our growing data platform team to develop data
infrastructure to support our next generation products and data initiatives. In this group, you will be
responsible for building, operationalizing and optimizing our data and data pipeline architecture, as
well as optimizing data flow and collection for cross functional teams. You should be a skillful data
pipeline builder and data wrangler who has experience of optimizing data systems and building
them from the ground up. You are expected to support our software developers, database
architects, data analysts and data scientists on data initiatives and ensure optimal data delivery
architecture is consistent throughout ongoing projects.

You must have excellent communication skills, have a strong sense of ownership, be self-directed
and comfortable supporting the data needs of multiple cross-functional teams.

This is a hands-on role and requires prior programming experience working with data on cloud.

Responsibilities

• Data Pipeline Building, Optimization and Operationalization


• Create and maintain optimal data pipeline architecture,
• Assemble large, complex data sets that meet functional / non-functional business
requirements.
• Identify, design, and implement internal process improvements: automating manual
processes, optimizing data delivery, re-designing infrastructure for greater scalability,
• Operationalizing, Improving the quality and reliability of data pipelines (monitoring, retry,
failure detection) Data Platform and Self-serving Data Infrastructure
• Build the infrastructure required for optimal extraction, transformation, and loading of data
from a wide variety of data sources using various data tools and cloud services.
• Design and build abstractions that hide the complexity of the underlying big data processing
and execution stack and that allow data consumers/partners to focus on their strengths:
product, data modeling, data analysis, search, information retrieval, and machine learning.
• Build processes and infrastructure to manage lifecycle of datasets: data structures,
metadata, dependency and workload management.
• Building internal services and tools to help in-house partners implement, deploy and analyze
datasets with a high-level of autonomy and limited friction.

Qualifications

• MS or BS in CS, Engineering, Math, Statistics, or a related field or equivalent practical


experience in data engineering.
• 3-5 years of experience working in data engineering using Python on Cloud platforms like
AWS. Proficient in programming in Python. With experience of S3, Glue, Snowflake, REST API
• Expert in SQL programming and experience working with relational databases, as well as
working familiarity with Redshift.
• Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
• Strong analytic skills related to working with unstructured datasets. A successful history of
manipulating, processing and extracting value from large, disconnected datasets.
• Strong problem-solving skills to understand complex structured and/or unstructured
datasets and their relationships to find issues and provide programmatic solutions.
• Working experience in data analytics (data wrangling, integration, visualization), and
reporting using BI (Business Intelligence) tools
• Experience performing root cause analysis on internal and external data and processes to
answer specific business questions and identify opportunities for improvement.
• Strong project management and organizational skills. Strong sense of ownership.
• Collaborative personality with experience supporting and working with cross-functional
teams in a dynamic environment.

Extra

• Proficiency in other high-level programming languages (Go, Java, Scala, or equivalent)


• Experience with Technologies like Kubeflow, EKS, Docker
• Experience with machine learning algorithms
• Data-driven approach to problem solving
• The ability to visualize and communicate complex concepts

You might also like