San Isidro, Buenos Aires
We are looking for DataBrinks Developer/ Architect who responsibilities will be:
- Designing and implementing highly performance data ingestion pipelines from multiple sources using Apache Spark and/or Azure Databricks
- Delivering and presenting proofs of concept to of key technology components to project stakeholders.
- Developing scalable and re-usable frameworks
- Integrating the end to end data pipeline to take data from source systems to target data repositories ensuring the quality and consistency of data is maintained at all times
- Working with event based / streaming technologies to ingest and process data
- Working with other members of the project team to support delivery of additional project components (API interfaces, Search)
- Evaluating the performance and applicability of multiple tools against customer requirements
- Working within an Agile delivery / DevOps methodology to deliver proof of concept and production implementation in iterative sprints.
Requisitos
- Strong knowledge of Data Management principles
- Experience in building ETL / data warehouse transformation processes
- Direct experience of building data piplines using Azure Data Factory and Apache Spark (preferably Databricks).
- Microsoft Azure Big Data Architecture certification (Desired but not required).
- Hands on experience designing and delivering solutions using the Azure Data Analytics platform (Cortana Intelligence Platform) including Azure Storage, Azure SQL Data Warehouse, Azure Data Lake, Azure Cosmos DB, Azure Stream Analytics
- Experience with Apache Kafka / Nifi for use with streaming data / event-based data
- Experience with other Open Source big data products Hadoop (incl. Hive, Pig, Impala)
- Experience with Open Source non-relational / NoSQL data repositories (incl. MongoDB, Cassandra, Neo4J)
- Experience working with structured and unstructured data
- Experience working in a Dev/Ops environment with tools such as Microsoft Visual Studio Team Services, Chef, Puppet or Terraform
Beneficios