Hadoop Engineer - Pleasanton, United States - Saicon Consultants, Inc

    Default job background
    Description

    Job Description

    Job Description

    TECHNICALKNOWLEDGE AND SKILLS:

    Consultantresources shall possess most of the following technical knowledgeandexperience:

    • Provide technical leadership,develop vision, gather requirements and translate client user requirements intotechnical architecture.
    • Strong Hands-on Experience inbuilding, deploying and productionizing ML models using software such as SparkMLLib, TensorFlow, PyTorch, Python Scikit-learn etc. is mandatory
    • Ability to evaluate and choosebest suited ML algorithms, perform feature engineering and optimize MachineLearning Models is mandatory
    • Strong fundamentals inalgorithms, data structures, statistics, predictive modeling, & distributedsystems is must
    • Design and implement an integratedBig Data platform and analytics solution
    • Design and implement datacollectors to collect and transport data to the Big Data Platform.
    • 4+ years of hands-onDevelopment, Deployment and production Support experience in Hadoopenvironment.
    • 4-5 years of programmingexperience in Java, Scala, Python.
    • Proficient in SQL andrelational database design and methods for data retrieval.
    • Knowledge of NoSQL systemslike HBase or Cassandra
    • Hands-on experience inCloudera Distribution 5.x
    • Hands-on experience increating, indexing Solr collections in Solr Cloud environment.
    • Hands-on experience buildingdata pipelines using Hadoop components Sqoop, Hive, Pig, Solr, MR, Spark, SparkSQL.
    • Must have experience withdeveloping Hive QL, UDF's for analyzing semi structured/structured datasets.
    • Must have experience withSpring framework
    • Hands-on experience ingestingand processing various file formats like Avro/Parquet/Sequence Files/Text Filesetc.
    • Hands-on experience working inReal-Time analytics like Spark/Kafka/Storm
    • Experience with GraphDatabases like Neo4J, Tiger Graph, Orient DB
    • Must have working experiencein the data warehousing and Business Intelligence systems.
    • Expertise in Unix/Linuxenvironment in writing scripts and schedule/execute jobs.
    • Successful track record ofbuilding automation scripts/code using Java, Bash, Python etc. and experiencein production support issue resolution process.
    • Experience with R,Jupyter/Zeppelin

    PREFERREDSKILLS:

    • Strong SQL skills
    • Java, Spring, Scala, ClouderaHadoop, MLLib, Spark, HBase, Neo4j, Solr, Python, Machine Learning