What will you be doing?
Working with the Data Architects to implement data pipelines
Working with our Big Data Principal Architects in the development both proof of concepts and complete implementations.
Working on complex and varied Big Data projects including tasks such as collecting, parsing, managing, analyzing, and visualizing very large datasets.
Translating complex functional and technical requirements into detailed designs.
Writing high-performance, reliable and maintainable code.
Performing data processing requirements analysis.
Performance tuning for batch and real-time data processing.
Securing components of clients’ Big Data platforms.
Diagnostics and troubleshooting of operational issues.
Health-checks and configuration reviews.
Data pipelines development ingestion, transformation, cleansing.
Data flow integration with external systems.
Integration with data access tools and products.
Assisting application developers and advising on efficient data access and manipulations.
Defining and implementing efficient operational processes
Dogfooding our product is important, so a short ( 3 month) on call rotation with second level support is required for all team members.
This works out to needing to be on call less than once a year and ensures quality of implementations.
What do we need from you?
While we realise you might not have everything on the list to be the successful candidate for the Big Data Developer job you will likely have at least 5 years experience in similar roles.
The position requires specialized knowledge and experience in performing the following :
Experience building data pipelines in any public cloud (GCP Dataflow, Glue, Azure DataFactory) or any equivalent
Experience writing ETL (Any popular tools)
Experience in data modeling, data design and persistence (e.g. warehousing, data marts, data lakes).
Strong Knowledge of Big Data architectures and distributed data processing frameworks : Hadoop, Spark, Kafka, Hive
Experience and working knowledge of various development platforms, frameworks and languages such as Java, Python, Scala and SQL
Experience with Apache Airflow, Oozie and Nifi would be great
General knowledge of modern data-center and cloud infrastructure including server hardware, networking and storage.
Strong written and verbal English communication skills
Experience with BI platforms, reporting tools, data visualization products, ETL engines.
Experience with data streaming frameworks.
DevOps experience with a good understanding of continuous delivery and deployment patterns and tools (Jenkins, Artifactory, Maven, etc)
Experience with Hbase.
Experience in data management best practices, real-time and batch data integration, and data rationalization
What do you get in return?
Competitive total rewards package
Flexible work environment : Why commute? Work remotely from your home, there’s no daily travel requirement to an office!
Outstanding people : Collaborate with the industry’s top minds.
Substantial training allowance : Hone your skills or learn new ones; participate in professional development days, attend conferences, become certified, whatever you like!
Amazing time off : Start with a minimum 3 weeks vacation, 7 sick days, and 2 professional development days!
Office Allowance : A device of your choosing and personalise your work environment!
Fun, fun, fun : Blog during work hours; take a day off and volunteer for your favorite charity.