- Expertise in processing huge size of data using Spark & ETL tools (Alteryx or any top ETL tool)
- Can perform complex transformations & aggregations required for BI dashboards
- Good knowledge of RDBMS & expert in SQL & writing materialized views/cubes etc.
- Knowledge & experience in Snowflake cloud MPP warehouse would be preferred
- Should have knowledge of AWS services & implemented data processing on cloud (AWS)
- Assemble large, complex data sets that meet functional / non-functional business requirements.
- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS ‘big data’ technologies.
- Must be good hands processing data in spark using Scala or Pyspark
- Expert in Natural Language Processing & text mining. Should have implemented various techniques to extract features from unstructured data (text/log/pdf etc.) and use these features to solve problems like sentiment analysis, topic modeling
- Feature Extraction like Part of speech tag, Bag of words, N grams, TF-IDF, Word2Vec, word embedding, etc.
- Good knowledge & experience in Knowledge Graph using unstructured data would be added advantage
- Knowledge of Graph DB like Neo4j, Datastax, or any Graph DB is added advantage.
- Experience in python is must
- Should have performed mapping of source to target in one of the ETL implementation
- Good problem-solving skills & quick grasping to understand functionality & implement the same technically.
There are multiple openings at Junior and Senior levels.