Data Engineer – Hybrid

Job Title: Data Engineer
Location: Charlotte, NC. Hybrid-3 days/week on site in Charlotte. Must be local or willing to relocate before starting.
Rate: $75/hr
Visa: USC, GC
Duration:12 Months Hybrid Schedule

Job Summary:
• Design and implement automated spark-based framework to facilitate data ingestion, transformation and consumption.
• Implement security protocols such as Kerberos Authentication, Encryption of data at rest, data authorization mechanism such as role-based access control using Apache ranger.
• Design and develop automated testing framework to perform data validation.
• Enhance existing spark-based frameworks to overcome tool limitations, and/or to add more features based on consumer expectations.
• Design and build high performing and scalable data pipeline platform using Hadoop, Apache Spark, MongoDB, Kafka and object storage architecture.
• Work with Infrastructure Engineers and System Administrators as appropriate in designing the big-data infrastructure
• Collaborate with application partners, Architects, Data Analysts and Modelers to build scalable and performant data solutions.
• Effectively work in a hybrid environment where legacy ETL and Data Warehouse applications and new big-data applications co-exist
• Work with Infrastructure Engineers and System Administrators as appropriate in designing the big-data infrastructure.
• Support ongoing data management efforts for Development, QA and Production environments
• Provide tool support, help consumers troubleshooting pipeline issues.
• Utilizes a thorough understanding of available technology, tools, and existing designs.
• Leverage knowledge of industry trends to build best in class technology to provide competitive advantage.

Required Qualification:
• 5+ years of experience of software engineering experience
• 5+ years of experience delivering complex enterprise-wide information technology solutions
• 5+ years of experience delivering ETL, data warehouse and data analytics capabilities on big-data architecture such as Hadoop
• 5+ years of Apache Spark design and development experience using Scala, Java, Python or Data Frames with Resilient Distributed Datasets (RDDs), Parquet or ORC file formats
• 6+ years of ETL (Extract, Transform, Load) Programming experience
• 2+ years of Kafka or equivalent experience
• 2+ years of NoSQL DB like Couchbase/MongoDB experience.
• 5+ experience working with complex SQLs and performance tuning

Interested candidates email your resume to chandra@amtexenterprises.com & shabu@amtexenterprises.com

To apply for this job email your details to chandra@amtexenterprises.com