João Victor

Latin Americas
Brazil

$65

English
About me

With Brazilian and Portuguese nationality (EU citizen), I currently work as a data engineer, with experiences in business intelligence and data science as well. In my academic and professional experiences, I could improve my autodidact and analytical skills and my profile as ”problem solver”.


During my experiences I was able to work on different projects covering the following topics:

- Development, maintenance and monitoring of data pipelines mainly using Python, SQL, Spark for data processing, cloud AWS, bash/shell, relational and non‐relational databases, Git for version control and Airflow as a orchestrator;

- Integrations mainly using Python with APIs;

- Development of Flask APIs;

- Data visualization using mainly Metabase and Tableau tools, as well as some Python libraries like Plotly, Seaborn, Matplotlib.

Skills
Data Engineering
100.0%
(5yrs)
Skill Python Python
90.0%
(7yrs)
Data Bricks
80.0%
(2.5yrs)
DBT
80.0%
(1yrs)
Spark
90.0%
(3.5yrs)
Data Warehousing
90.0%
(7yrs)
Skill AWS AWS
80.0%
(7yrs)
Airflow
80.0%
(6yrs)
Skill Azure Azure
70.0%
(2yrs)
SQL
90.0%
(7yrs)
Experience
Data Engineer | Tenda Construction Company · Full-time
Apr 2024 - Present

• Development, maintenance, monitoring and orchestration of data pipelines on AWS using Python, Pentaho Data Integration and DBT

• Data Modeling in Redshift Data Warehouse / S3 Data Lake

• Data pipeline orchestration and CI/CD using Rundeck

• Productization of Data Science team models

• Development of app using Streamlit

• Data Architecture

Data EngineerData Engineer | Shape
Sep 2021 - Jun 2023

• Responsible for the development, maintenance and monitoring of data pipelines for stakeholders in Asia using mainly Azure resources like Azure Databricks (Spark) for data processing, Blob Storage, SQL Server

• Data pipeline orchestration using Azure Data Factory, Airflow and Prefect

• Development of Flask APIs, which are the core of the main client application

• Database versioning using Liquibase

• CI/CD using Azure DevOps/Pipelines

Data Engineer | Infoprice
Nov 2020 - Sep 2021

• Development, maintenance and monitoring of data pipelines mainly using the programming languages Python and SQL, data processing using Apache Hive, Apache Hadoop and Apache Spark (PySpark), cloud services in AWS (S3, RDS, EMR and EC2), bash/shell, relational and non‑relational databases, Git for version control and Airflow as a orchestrator.

• Integrations mainly using Python with APIs

• Development of Flask APIs

• Web Scrapper and Crawler development using Scrapy and Requests (Python)

• Data streaming using Apache Kafka

Data Intelligence Analyst | Intelipost
Mar 2020 - Nov 2020 ·

Using Python, Shell Script, AWS, Xplenty, Redshift, Postgres, SQL and Metabase for maintenance of the data structure, automation and optimization of internal processes, project development with the Data Visualization team, ETL and data enrichment.