Job Description
4 or more years of experience of data engineering and Big Data systems in a project services or product engineering organisation
Proficient understanding of distributed computing principles
Hands on experience in programming or scripting languages like Java, Javascript, Ruby, Python, R
Good knowledge of various ETL techniques and frameworks such as Flume
Experience with various messaging systems such as Kafka, RabbitMQ
Experience with building stream processing systems using solutions such as Storm, Spark Streaming, Kinesis
Experience on Data Visualisation frameworks like D3.js
Experience with NoSQL databases such as HBase, Cassandra, MongoDB
Good Knowledge of Big data querying tools such as Hive, Pig
Experience with Machine Learning toolkits such as Mahout, SparkML, H2O
Good understanding of Lambda architecture and when to use them
Experience or Knowledge on Columnar Databases such as Redshift
Experience or Knowledge on Deep learning frameworks like Torch, Caffe, TensorFlow would give you an edge.
Ability to manage Hadoop clusters and to solve any ongoing issues with operating the clusters
An engineering Degree in computer science or equivalent practical experience
Prior experience working within the Media, E-commerce, Telecommunications or similar large scale consumer facing industry