Job Description

  • Expertise in processing huge size of data using Spark & ETL tools (Alteryx or any top ETL tool)
  • Can perform complex transformations & aggregations required for BI dashboards
  • Good knowledge of RDBMS & expert in SQL & writing materialized views/cubes etc.
  • Knowledge & experience in Snowflake cloud MPP warehouse would be preferred
  • Should have knowledge of AWS services & implemented data processing on cloud (AWS)
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS ‘big data’ technologies.
  • Must be good hands processing data in spark using Scala or Pyspark
  • Expert in Natural Language Processing & text mining. Should have implemented various techniques to extract features from unstructured data (text/log/pdf etc.) and use these features to solve problems like sentiment analysis, topic modeling
  • Feature Extraction like Part of speech tag, Bag of words, N grams, TF-IDF, Word2Vec, word embedding, etc.
  • Good knowledge & experience in Knowledge Graph using unstructured data would be added advantage
  • Knowledge of Graph DB like Neo4j, Datastax, or any Graph DB is added advantage.
  • Experience in python is must
  • Should have performed mapping of source to target in one of the ETL implementation
  • Good problem-solving skills & quick grasping to understand functionality & implement the same technically.
There are multiple openings at Junior and Senior levels.

Skills

Big Data
Spark
ETL
Knowledge Graph

Response Form