Requirements
So what is exactly needed for this position? Well, first you must have at least 5 - 6 years of broad information technology experience. What else? Strong knowledge of Data Warehousing concepts with understanding of traditional and MPP database designs, star and snowflake schemas with at least 4-5 years of data modeling. You must have at least 2 years of hands-on development experience using ETL Tools such as Pentaho, SSIS, Informatica, Talend, Fivetran or Airflow. Knowledge on the architecture, design and implementation of MPP Databases such as Teradata, Snowflake or Redshift. 2-year development using Cloud-based analytics solutions preferrable AWS or GCP. Knowledge of designing and implementing streaming pipelines using Apache Kafka, Apache Spark, or Segment. At least 2 years experience of using Python in a cloud based environment is definitely a plus. Knowledge on NoSQL DBs is not required but preferred.
We need someone who has a Bachelor’s degree in Computer Science (have an MBA or MS degree? That’s a plus!) or 8-10 years experience in lieu of degree. We need someone who has knowledge of the following: Python, Apache Kafka, AWS Redshift, AWS Glue, AWS S3, Pentaho Data Integration.
We need someone who has knowledge in Database Tuning and ETL tuning. We need someone who has the ability to work effectively across internal functional areas in ambiguous situations. Someone who is a structured thinker and effective communicator.