Our client, the World’s leading provider of premium products and premium mobility seeks a SENIOR DATA [URL Removed] company goes beyond this expectation to bring first-class cars, and that explains why the German corporation is held in high ostentation across the globe. Understanding the critical regional business landscapes, and this plays a major role in its Global stability. Minimum requirements:

  • 10+ years extensive experience within the data science industry in a fast-paced and complex business setting and with leading teams
  • Extensive experience in Machine Learning solving business problems
  • Agile working experience advantageous

Role Tasks:

  • Strong working knowledge with software development tools, techniques and approaches used to build application solutions
  • Implements analytical algorithms for object detection, segmentation, classification, and recognition
  • Is responsible for discovering insights and identifying business opportunities using algorithmic, statistical, and mining techniques
  • Plays a strategic role in creative ideas for leveraging the business’ vast collection of data into new ground-breaking IT solutions
  • Establishes accurate and scalable analytics systems
  • Interprets the results of statistical and predictive experiments and regression analysis and integrates this into complex business processes
  • Conducts customer specification studies, gathers requirements, conducts system architectural design and turns the requirements into a final product
  • Provides complete application lifecycle development, deployment, and operations support for Big Data solutions and infrastructure
  • Collaborates with various team members and facilitates the development, automation, and seamless delivery of analytics solutions into Big Data clusters
  • Is responsible for Data Warehousing and ETL using Informatica Power Center
  • Analyses the data and build reports by using data visualisation tools such as Tableau, Cognos, MicroStrategy
  • Imports and exports data using Sqoop from HDFS to Relational Database Systems and vice-versa
  • Codes and tests the Standardisation, Normalisation, Load, Extract and AVRO models to filter / massage the data and its validation
  • Installs, configures, and uses ecosystem components like Hadoop Map Reduce, Spark, Hive, Sqoop, Pig, HDFS, HBase, Cassandra, ZooKeeper, Oozie, Hue, Impala and Flume.

Technical skills:

  • Strong background in mathematics
  • Good analytical and problem-solving skills
  • Assisting with the business case
  • Requirement’s organisation
  • Translating and simplifying requirements
  • Document requirements in appropriate format
  • depending on methodology followed
  • Assist with identification and management of risks
  • Understanding of dimensional modelling and building ETL pipelines
  • Familiarity with Tableau, QlikView, Mode, or similar data visualization tools

Languages Python, Scala, SQL, Java, PL/SQL

Web Technologies

Web Service, SOAP, Rest web services, JSP

Big Data Eco System Spark, HDFS, Yarn, Map Reduce,

Hive, Pig, Sqoop, ZooKeeper, Kafka,

Oozie, Hue, Impala, Flume

Scripting

Languages HTML, JavaScript, CSS, XML and

Ajax

Machine

Learning

R, SAS, SKLearn, MATLAB, Octave,

Spark ML

No SQL

Databases Cassandra, HBase, MongoDB,

Vertica

Cloud AWS, Azure Cloud

Operating System Windows, Linux and Unix

BI/DWH/ETL

Tools Informatica 9.5/9.1/8.6, Tableau,

Cognos

DBMS / RDBMS Oracle 12c/11g, SQL Server 2014,

DB2, Teradata 14/12, AWS Redshift

IDEs Eclipse, Jupiter Notebooks,

Microsoft Visual Studio, Flex

Builder, Spyder, TOAD, NetBeans,

PL/SQL Developer, Putty, Squirrel

SQL

Version Control SVN, CVS, Git, and Rational Clear

Case

Tools

  • FileZilla, JUnit, Splunk, Clear Quest,

Rally, Jira, Confluence, Bitbucket

Desired Skills:

  • Data Scientist
  • machine learning

Learn more/Apply for this position