The Smart Cube : Assistant Manager – Data Engineer

Brief Description of position:

About us

In today’s Intelligence Age, the world’s best businesses have no choice but to continuously strive to become smarter.

Our clients hold the key to this and are today harnessing the power of The Smart Cube’s solutions to unlock new innovation and profit streams.

The strength of our solutions is born of a powerful blend of people, tools and technology.

A rare breed

Back in 2003, the world was a very different place - the consumer internet was less than 10 years old and the smartphone was still a novelty.

The Smart Cube was created as a new kind of business, one that would make sense of the reams of data being generated.

Since then it’s been quite a journey.

A journey that has seen us change the way organisations think about intelligence. A journey that has put our industry on the map. A journey that is only getting started.

Our future focus

We do not exist to help our clients simply fire-fight today. We help them to look ahead and stay ahead. Our vision is to be the catalyst for organisational intelligence for the world’s leading firms.

What job we expect you to do:

  • Connecting, designing, scheduling, and deploying data warehouse systems
  • Developing data pipelines and enable dash boards for stakeholders and
  • Develop, construct, test and maintain system architectures
  • Create best practices for data loading and extraction
  • Doing quick POCs for any data eccentric development task

What skills we expect you to bring:

  • Strong programing skills, being well versed in Object-Oriented Programming system (OOPS), data structures, and algorithms
  • Should be comfortable in executing ETL (Extract, Transform and Load) processes which include data ingestion, data cleaning and curation into a data warehouse, database, or data platform
  • Should be comfortable with schema designing
  • Experience in distributed computing environment
  • Experience in structured/unstructured data and batch processing/real-time processing (good to have)
  • Be comfortable with SQL (mandatory), Python(mandatory), Scala (good to have) to manipulate and prepare data and conduct various analysis as needed
  • Reading\writing data to\from various sources - APIs, cloud storage, databases, big data platforms
  • Experience of working with Big Data environment such as Hadoop and the ecosystem
  • Data transformations and applying ML models
  • Creating web services to allow create, read, update and delete (CRUD) operations
  • Competent in project management framework such as Agile
  • Excellent communication skills, both written and verbal

What expertise will bring added advantage?

  • Machine learning
  • Statistical Modelling
  • Natural Language Processing

What Tools and Technologies we expect you to know? We understand one cannot be master of all.

  • Python - pandas, django\flask, sklearn, scikit
  • SQL, BigQuery
  • Hadoop ecosystems (HDFS, HIVE, Mapreduce, Pig, Spark, Hadoop etc.)
  • Kafka
  • Apache Spark
  • Linux
  • Airflow

How many years of experience do you need?

Minimum 4 years of relevant experience

Where is the job location?

We are ready to welcome you at Noida and Gurgaon and also open for any other location in India

Location
Gurgaon,Noida
Support

We use cookies essential for this site to function well. Please click to help us improve its usefulness with additional cookies. Learn about our use of cookies in our Privacy Policy.

Feedback

We believe in making Analytics Vidhya the best experience possible for Data Science enthusiasts. Help us by providing valuable Feedback.