The Senior Data Engineer will design and implement scalable data integration solutions using Databricks, Apache Spark, and cloud-based platforms, collaborating with cross-functional teams to deliver high-quality solutions that meet business requirements.
Requirements
- 5+ years of experience with Databricks, Apache Spark, and cloud-based data integration
- Strong Technical expertise with cloud-based platforms, including AWS and or Azure cloud
- Strong programming skills in languages like SQL, Python, Java, or Scala
- 3+ years' experience with cloud-based data infrastructure and integration leveraging tools like S3, Airflow, EC2, AWS Glue, DynamoDB & Lambdas, Athena, AWS Code deploy, Azure Data Factory, or Google Cloud Dataflow
- Experience with Jenkins and other CI/CD tools like GitLab CI/CD, CircleCI, etc.
- Experience with containerization using Docker and Kubernetes
- Experience with infrastructure such as code using tools like Terraform or CloudFormation
- Experience with Agile development methodologies and version control systems like Git
- Experience with IT service management tools like ServiceNow, JIRA, etc.
- Data warehousing solutions, such as Amazon Redshift, Azure Synapse Analytics, or Google BigQuery will be a plus but not mandatory.
- Data science and machine learning concepts, including TensorFlow, PyTorch, or scikit-learn will be a plus but not mandatory.
- Strong technical background in computer science, software engineering, or a related field.
- Excellent collaboration, communication, and interpersonal skills.
- Experience with data governance, data quality, and data security principles.
- Ability to lead and mentor junior team members.
- AWS Certified Solutions Architect or AWS Certified Developer Associate or Azure Certified Solutions Architect certification.