Job D5677 Description Click to apply: Please attach resume to mail
SOFT's client located in New York, NY ( Hybrid ) is looking for a Data Engineer - Data Bricks / AWS for a long term contract assignment.

Qualifications:  
• Minimum of 5 years of experience in data engineering roles, with a focus on AWS and Databricks.  
• Highly proficient with Databricks, Spark, Starburst/Trino, Python, PySpark and SQL  
• Hands-on experience in Gitlab with CI/CD.  
• Hands-on experience in AWS Services like S3, RDS, Lambda, SQS, SNS, MSK is required.  
• Strong SQL skills to perform data analysis and understanding of source data.  
• Experience with data pipeline orchestration tools  

Responsibilities:
Design, develop, monitor, and maintain data pipelines in an AWS ecosystem with Databricks, Delta Lake, Python, SQL and Starburst as the technology stack. Collaborate with cross-functional teams to understand data needs and translate them into effective data pipeline solutions.  
• Establish data quality checks and ensure data integrity and accuracy throughout the data lifecycle.  
• Automate testing of the data pipelines and configure as part of CICD  
• Optimize data processing and query performance for large-scale datasets within AWS and Databricks environments.  
• Document data engineering processes, architecture, and configurations.  
• Troubleshooting and debugging data-related issues on the AWS Databricks platform.  
• Integrating Databricks with other AWS products such as SNS, SQS, and MSK.  
Comments/Special Instructions