4-6 years of experience:
Experience in hadoop/big data technologies.
experience in spark.
experience in Snowflake
experience working on Google or AWS cloud developing data solutions. Certifications preferred.
Hands-on experience with Python/Pyspark/Scala and basic libraries for machine learning is required;
Experience with containerization and related technologies (e.g. Docker, Kubernetes)
Experience with all aspects of DevOps (source control, continuous integration, deployments, etc.)
Comprehensive knowledge of the principles of software engineering and data analytics
Advanced knowledge of the Hadoop ecosystem and Big Data technologies Hands-on experience with the Hadoop eco-system (HDFS, MapReduce, Hive, Pig, Impala, Spark, Kafka, Kudu, Solr)
Knowledge of agile (scrum) development methodology is a plus
Strong development/automation skills
Proficient in programming in Java or Python with prior Apache Beam/Spark experience a plus.
System level understanding - Data structures, algorithms, distributed storage & compute
Can-do attitude on solving complex business problems, good interpersonal and teamwork skills