Description

Our client represents the connected world, offering innovative and customer-centric information technology experiences, enabling Enterprises, Associates, and Society to Rise™.

They are a USD 6 billion company with 163,000+ professionals across 90 countries, helping 1279 global customers, including Fortune 500 companies. They focus on leveraging next-generation technologies, including 5G, Blockchain, Metaverse, Quantum Computing, Cybersecurity, Artificial Intelligence, and more, on enabling end-to-end digital transformation for global customers.

Our client is one of the fastest-growing brands and among the top 7 IT service providers globally. Our client has consistently emerged as a leader in sustainability and is recognized amongst the ‘2021 Global 100 Most sustainable corporations in the World by Corporate Knights. 

We are currently searching for a Databricks Architect:

Responsibilities:

  • Design and develop modern cloud data platforms.
  • Perform complex coding using SQL, Spark, Python, and T-SQL.
  • Develop and maintain solutions using Azure stack (Azure Data Lake, Azure Data Factory, and Databricks).
  • Implement ETL processes using tools such as Talend, SSIS, and Informatica.
  • Build data analytics solutions with a focus on performance and scalability.
  • Implement data warehousing solutions.
  • Work as a Data Engineer within an Azure Big Data environment.
  • Develop and optimize data models and data engineering solutions.
  • Utilize Scala or Python programming for data processing.
  • Leverage Azure services such as Azure Data Lake Analytics, U-SQL, and Azure SQL Data Warehouse.
  • Apply analytical and problem-solving skills in a big data environment.
  • Utilize Lambda Architecture and data warehousing concepts.
  • Manage source code using version control systems like GIT.
  • Deploy Azure resources using Azure ARM Templates.

Requirements:

  • Bachelor's degree in Computer Science, Engineering, IT, Mathematics, or a related field.
  • 8 to 10 years of experience in a Data Engineering role.
  • Strong background in Data Science, Data Engineering, and automation within cloud-based global platforms.
  • Experience with MS SQL, Azure Data Factory, Docker, and containerization tools.
  • Expertise in cataloging and governance solutions for streaming with Databricks Delta Live Tables.
  • Experience with multi-geo, multi-tier service design and operations.
  • Proficiency in modern delivery methodologies such as SAFe Agile, Iterative, and Waterfall.
  • Strong verbal and written communication skills in English.
  • Excellent interpersonal and organizational skills for collaborating in distributed global teams.
  • Ability to break down complex technical challenges into measurable and explainable decisions.
  • Willingness to learn and adopt emerging technologies, defining engineering standards and automation processes.
  • Operational experience, including early life support and root cause analysis.

Languages:

  • Advanced oral English.
  • Native Spanish.

Notes:

  • Fully remote.

If you meet these qualifications and are pursuing new challenges, Start your application to join an award-winning employer. Explore all our job openings | Sequoia Career’s Page: https://www.sequoia-connect.com/careers/.

Position image
Map showing company location