Job Description:
Ensuring high quality software development
with complete documentation and traceability
Promoting unit/functional testing and code inspection processes
Working with business partners and BA’s to understand requirements
Working with data scientists to understand and interpret complex datasets
Assisting with the provision of task breakdown and estimates
Qualifications:
5+ Hands-on experience in Hadoop
Spark
Hive
Kafka and Unix Scripting
4+ Experience with relational SQL and NoSQL databases: Oracle
MongoDB
HBase
3+ Experience with Java/Python/Scala and SQL.
Experience in CDH
Experienced in working with large and multiple datasets and data warehouses
Experience building and optimizing ‘big data’ data pipelines
architectures and data sets.
Strong analytic skills and experience working with unstructured datasets
Ability to effectively use complex analytical
interpretive and problem solving techniques
Demonstrated interpersonal
verbal and written communication skills
Good to Have:
Experience with Confluent Kafka
Experience with Redhat JBPM
Integration with search solution to middleware & distributed messaging - Kafka
Tibco EMS
Experience with CI/CD build pipelines and toolchain – Git
BitBucket
Jira
Experience with external cloud platform such as AWS
Experience with container technologies (Docker
Pivotal Cloud Foundry) and supporting frameworks (Kubernetes
OpenShift
Mesos)
Education:
Bachelor’s/University degree or equivalent experience