Data Lake Cloud Engineer (GTH)Country: Mexico
- Bachelor’s degree in computer science of related degree
- Previous experience working with AWS Data Lake technologies.
- Knowledge of programming languages such as Python, Java, or Scala.
- Familiarity with ETL tools and data processing platforms like Apache Spark or Hadoop.
- Solid understanding of SQL and NoSQL databases.
- Fundamental AWS Knowledge:Basic AWS Services: Familiarity with key AWS services such as EC (Elastic Compute Cloud), S (Simple Storage Service), RDS (Relational Database Service), VPC (Virtual Private Cloud), IAM (Identity and Access Management), and Lambda., Cloudformation, DMS, Glue, EMR; Kinesis, Iceberg AWSAWS Architecture: Understanding the principles of AWS architecture and design patterns to build scalable and secure solutions.
- Infrastructure as Code (IaC)IaC Tools: Practical knowledge of tools like AWS CloudFormation, Terraform, and AWS CLI.
This includes the ability to write and maintain IaC templates to define and manage infrastructure.
- Configuration Management and AutomationAWS Systems Manager: Familiarity with AWS Systems Manager for automating tasks, managing configurations, and administering instances.Automation Scripts: Ability to write scripts in Bash, Python, or PowerShell to automate administrative and configuration tasks.Knowledge of configuration management systems, such as Ansible or related systems.
- Networking and SecurityNetworking Fundamentals: Understanding basic networking concepts such as subnets, IP addresses, routing, and gateways, as well as configuring VPCs and subnets in AWS.AWS Security: Knowledge of identity and access management (IAM) and best security practices to protect infrastructure and data.
This includes using policies, roles, and groups.
- Monitoring and MaintenanceAWS CloudWatch: Ability to configure and use Amazon CloudWatch to monitor resources, create alarms, and generate logs.AWS CloudTrail: Understanding how to use AWS CloudTrail to audit activity in the account and ensure compliance with policies.Experience in monitoring frameworks such as Splunk, Grafana, Dynatrace, or similar.
- Database ManagementRDS and DynamoDB: Knowledge of AWS databases, including Amazon RDS for relational databases and Amazon DynamoDB for NoSQL databases.Backups and Recovery: Understanding backup and recovery strategies for cloud databases.
- Implementation and OperationSoftware Development Life Cycle (SDLC): Familiarity with SDLC principles, including Continuous Integration (CI) and Continuous Deployment (CD) to implement changes efficiently.Change Management and Versioning: Understanding change management practices and using version control systems like Git to manage IaC code.Experience in three-tier architectures: web (Apache), Apache Spark, engine (Tomcat/Spring) and databases (MySQL, PostgreSQL, ORACLE and NoSQL databases).
- Troubleshooting and SupportDiagnosis and Troubleshooting: Skills to diagnose problems and perform troubleshooting on deployed infrastructure.Documentation: Ability to document configurations, processes, and changes made to the infrastructure.
- Best Practices and Architecture PatternsResilience Patterns: Understanding architectural patterns like high availability, disaster recovery, and scalability.Cost Optimization: Knowledge of practices to optimize costs when deploying and operating resources in AWS
- Strong experience in systems automation and Linux administration
Required Skill Profession
Computer Occupations
📌 Data Lake Cloud Engineer (GTH)
🏢 Santander
📍 Santiago de Querétaro