Vacancy caducado!
Your Opportunity
Do you want to be part of a Data Warehouse team handling over 120 terabytes of data and building the next generation analytics platform for a leading financial firm? At Schwab, the Global Data Technology (GDT) organization governs the strategy and implementation of the enterprise data warehouse and emerging data platforms. We help Sales, Marketing, Finance and executive leadership make fact-based decisions by integrating and analyzing data.We are looking for a data engineer who has passion for data integration technologies and comes with data warehousing background. Someone who has experience in designing and coding ETL and one who wants to be part of a team that is actively delivering projects in Teradata, Big Data and working towards migrating to cloud technology.What you are good at- Creating/updating ETL specifications and supporting documentation
- Developing ETL logic utilizing Informatica workflows, scripting and load utilities
- Building and maintaining code for big data ingestion using Talend, Scoop, Hive etc
- Implementing data flow scripts using Unix /Sqoop / Hive QL / Pig scripting
- Designing schemas, data models and data architecture for Hadoop and HBase environments
- Designing, building and support data processing pipelines to transform data in Big Data or Teradata platforms
- Developing and executing quality assurance and test scripts
- Work with business analysts to understand business requirements and use cases
- Problem solving and fixing technical issues
- Minimum of 3 years of experience in understanding of best practices for building and designing ETL code
- Strong SQL experience with the ability to develop, tune and debug complex SQL applications is required
- Knowledge in schema design, developing data models and proven ability to work with complex data is required
- Experience in ETL such as Informatica and Talend tools is required
- Strong Teradata coding and utilities experience is required
- Experience in working in large environments such as RDBMS, EDW, NoSQL, etc. is required
- Hands-on experience with Hadoop, MapReduce, Hive, SPARK, Kafka and HBASE is required
- Understanding Hadoop file format and compressions is required
- Experience with scheduling tools (eg. Control M, ESP)
- Understanding of best practices for building Data Lake and analytical architecture on Hadoop is preferred
- Scripting / programming with UNIX, Java, Python, Scala etc. is preferred
- Knowledge in real time data ingestion into Hadoop is preferred
- Experience with Test Driven Code Development, SCM tools such as GIT, Jenkins is preferred
- Experience collaborating with business and technology partners and offshore development teams
- Good interpersonal, analytical, problem-solving and organizational skills
- Excellent written/verbal communication skills
Vacancy caducado!