Buscas ser parte de una empresa innovadora, inclusiva y colaborativa, que valora tanto tu desarrollo profesional como el equilibrio entre tu vida personal y laboral, y que cuenta con diversidad de clientes líderes en varios sectores?
si es así, ¡descubre la vacante que deintec tiene para ti e impulsa tu carrera en it!
posición*:data engineer*
ubicación: cdmx, méxico
esquema: híbrido en oficinas de cdmx azcapotzalco
tipo de contrato: indeterminado
horario: lunes a viernes
idioma: inglés avanzado conversacional requerido.
c1, c2, nativo.
escolaridad: ingeniero computación, sistemas, etc.
*purpose of the position*
the data engineer will join the americas it software and architecture team.
this business centric it team allows for business-focused development of new, cutting edge, solutions as well as integrating overall global/regional it initiatives into the business.
in this role, we are seeking a talented and motivated data engineer with hands-on experience in apache spark and microsoft
fabric to join our dynamic data engineering team.
in this role, you will be at the heart of our data initiatives, designing robust data pipelines, optimizing data architectures, and enabling data-driven decision-making across the organization.
the data engineer would be expected to work independently and/or in conjunction with appropriate business unit/it leaders or cross-functional teams during any engagement.
they would be tasked with understanding technological and business problems and focused on delivering cutting-edge solutions.
the data engineer must possess the ability to architect or integrate the appropriate solution from the ground up based on a set of business requirements.
*principal responsibilities*
- design and develop data pipelines: architect, build, and maintain scalable data pipelines leveraging apache spark and microsoft fabric to process large volumes of structured and unstructured data from diverse sources.
- data integration: integrate data from internal and external sources, ensuring accuracy, consistency, and reliability throughout the data lifecycle.
- performance optimization: monitor and optimize etl/elt processes for performance, scalability, and cost-efficiency, proactively identifying bottlenecks and implementing improvements.
- data quality and governance: implement best practices for data quality, data cataloguing, and lineage, and support governance policies to ensure compliance and high standards.
- collaboration: work closely with data scientists, analysts, business stakeholders, and cross-functional it teams to gather requirements, deliver insights, and support analytical models.
- documentation: prepare and maintain detailed documentation for data workflows, pipeline architectures, data schemas, and processes.
- continuous improvement: stay current with emerging trends and technologies in data engineering, suggesting and implementing innovative solutions to enhance the data platform.
- troubleshooting: diagnose, debug, and resolve data pipeline and infrastructure issues to ensure robustness and reliability of the data ecosystem.
- security: adhere to and enforce data security and privacy guidelines, ensuring sensitive data is handled appropriately.
*required critical behaviours*
- excellent problem-solving skills and meticulous attention to detail.
- strong communication and teamwork abilities, with a collaborative approach to project deliver
*required key skills*
*mandatorias*
- proven experience working with apache spark for large-scale data processing (batch and streaming).
- hands-on expertise with microsoft fabric (formerly power bi dataflows, data factory within fabric, or similar cloud-based data integration platforms).
- strong proficiency in sql, python, and/or scala for data manipulation and etl development.
- solid understanding of distributed systems, data warehousing, and data modeling concepts.
- demonstrated ability to design, implement, and optimize etl/elt
- bachelor's or master's degree in computer science, engineering, information systems, or a related field.
- 2+ years of professional experience in data engineering, data architecture, or a related domain.
- experience with cloud platforms such as microsoft azure, aws, or google cloud, with a preference for azure.
- knowledge of data governance principles and data quality management.
*preferencial*
- knowledge of devops practices for data pipelines, including ci/cd, automated testing, and infrastructure as code.
- experience with real-time data processing and streaming frameworks (e.g., kafka, azure event hubs).
p
- professional certifications in apache spark, azure data engineering, or microsoft fabric/power platform.
- experience with data visualization and reporting tools (e.g., power bi, tableau).
*ofrecemos*:
- salario competitivo basado en experiência
- prestaciones superiores a las de ley (como: fondo de ahorro, caja de ahorro, sgmm, vales de despensa, etc.)
- oportunidad de cre