compensation:
$100K — $150K *
industry:
specialty:
experience:
RESPONSIBILITIES:
- Communicates and maintains Master Data, Metadata, Data Management Repositories, Logical Data Models, Data Standards
- Create and maintain optimal data pipeline architecture
- You will assemble large, complex data sets that meet functional / non-functional business requirements
- You will identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
- Build industrialized analytic datasets and delivery mechanisms that utilize the data pipeline to deliver actionable insights into customer acquisition, operational efficiency and other key business performance metrics
- Work with business partners on data-related technical issues and develop requirements to support their data infrastructure needs
- Create highly consistent and accurate analytic datasets suitable for business intelligence and data scientist team members
REQUIREMENTS:
- At least 3 years of hands on experience with Big Data Tools: Hadoop, Spark, Kafka, etc.
- You have mastery with databases - Advanced SQL and NoSQL databases, including Postgres and Cassandra
- Data Wrangling and Preparation: Alteryx, Trifacta, SAS, Datameer
- Stream-processing systems: Storm, Spark-Streaming, etc.
- 7 or more years with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
- Ability to tackle problems quickly and completely
- Ability to identify tasks which require automation and automate them
- A demonstrable understanding of networking/distributed computing environment concepts
- Ability to multi-task and stay organized in a dynamic work environment
PREFERRED:
- Data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
- AWS cloud services: EC2, EMR, RDS, Redshift
Valid through: 2/9/2021