*Please send me the profiles at* *[email protected]* <[email protected]>*m*
*Position : Data Lead Developer (AZURE)* *Location: Omaha, NE * *Duration: 6 months to start with (immediate position)* Responsibilities: • Designing and implementing highly performant data ingestion pipelines from multiple sources using Apache Spark and/or Azure Databricks ensuring that the infrastructure is highly-available and secure. • Develop scalable and re-usable frameworks for ingestion of data from centralized and federated hubs • Architect and build security compliant user management framework for multi-tenant big data platform • Performance-tune Apache Spark applications to optimize cluster configurations • Set up and maintain infrastructure for RESTful microservices layer • Integrate the end to end data pipeline to take data from source systems to target data repositories ensuring the quality and consistency of data is maintained at all times • Collaborate across internal functional groups and external vendors to incorporate emerging/novel big data tools • Help maintain and support the platform on a day-to-day basis including on-boarding new members • Work closely with the product management and development teams to rapidly translate the understanding of customer data and requirements to product and solutions • Enable ETL/ELT solutions and perform DB tuning, table partitioning, shell scripting, drive prototypes and POCs Qualifications: • Bachelor's Degree in Computer Science • Minimum 8 years of experience as a technology leader with strong knowledge of Data Management principles and dimensional modelling concepts • Advanced applied knowledge of Spark – a must. • Proven knowledge of designing, developing and deploying federated data architecture – highly desirable • Hands on experience designing and delivering solutions using with Microsoft Azure data services (Azure Data Lake, Azure Data Factory, Azure ML, Azure SQL, Azure Datawarehouse, Azure DataBricks, Azure Cosmos DB) -– highly desirable • Experience with event based / streaming technologies to ingest and process data (Apache KAFKA, Nifi, Azure Event Hub) – highly desirable • Experience with OS non-relational/NoSQL data repositories (incl. MongoDB, Cassandra, Neo4J) - desirable • Familiarity Open Source big data products Hadoop (incl. Hive, Pig, Impala) – desirable • Broad based architecture acumen: Database architecture, ETL, SOA, Cloud, etc - desirable • Ability to work successfully in a distributed team environment * Deepak Gulia *| Digitive LLC – cloud*.* made simple Fax: 408-935-8696 | Email: [email protected] *GTALK :- **[email protected]* <[email protected]> -- You received this message because you are subscribed to the Google Groups "Visa Transfer OPT, CPT, PT, H1,H4" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To view this discussion on the web visit https://groups.google.com/d/msgid/paul_talluri/CANidSO3Dg_d89RDsdpqH5yp%3D9q%2BST_9C-6qqhRG%3DvB8h9bjUtw%40mail.gmail.com.
