Vacancy caducado!
- Leading development of high-volume, low-latency, data solutions to support a variety of analytics use cases in an efficient and scalable manner
- Technical ownership of data platform roadmap and its successful delivery, including sources, flows, capabilities, and performance
- Partner with data analysts, data scientists, and functional stakeholders to define needs and continuously improve the data platform
- Manage internal and external teams of data engineers, data analysts, and program managers both internally and with independent software vendors
- Design and build mechanisms to manage streaming big data from a wide variety of sources
- Design and build mechanisms to tune, filter, and manage the flow of data across a distributed platform
- Define and evaluate key tasks and acceptance criteria of data systems in development and into production
- Contribute to the overall data platform architecture and evolution
- Normalize complicated data sources to convert potentially unusable data into a format that can be efficiently used by software and/or employees
- Proven track record of building and delivering large, highly-available, enterprise-grade data systems and solutions
- Master's degree in Computer Science, Computer Engineering, or a related technical degree; six years related experience; or equivalent combination of education and experience
- AWS: 3+ years experience with using a broad range of AWS technologies (e.g. EC2, EMR, S3, Lake Formation, Redshift, VPC, Glacier, IAM, CloudWatch, SQS, Lambda, CloudTrail, Systems Manager, KMS, Kinesis Streams)
- 3+ years experience in data streaming technologies, such as Kafka
- 3+ years experience and demonstrated strength in data modeling, data warehousing, ETL (Extract, Transform, and Load) development, and clear understanding of the difference and rationale of ELT versus ETL
- Ability to work in teams and collaborate with others to clarify requirements, quickly identify problems, and collaboratively find creative solutions
- Ability to assist in documenting requirements as well as resolve conflicts or ambiguities with a willingness to get his/her hands dirty and dive deep into technical details
- Prior experience building highly available, distributed systems as it pertains to big data storage, flow, and processing of large, complex data
- Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets
- Prior experience with vendor-specific solutions such as Confluent, Cloudera, Snowflake, etc.
- Working knowledge of modern data architectures and trends such as Kappa versus Lambda and Data Mesh
- Working knowledge of common data analyst/science workflows, business intelligence tools, and AI/ML modeling
- 2+ years experience in implementing data-driven solutions in a production environment using tools such as Hadoop, Impala, Hive, NiFi, Prometheus, Spark, Athena, Redshift, ElasticSearch, BigTable, or Airflow
- 2+ years experience with querying and designing databases using one or more of the following: MySQL, MS SQL, Oracle SQL, or other professional database system
Vacancy caducado!