Welcome! nice to
meet you.
I'm Jonathan Mota



About



Do

  • Working on projects in the Big Data area, using the current technologies PySpark, Apache Spark, Apache Kafka, Azure DataFactory, Databricks, Google Cloud Platform (GCP), Microsoft Azure.
  • I consider myself extremely dedicated, focused on goals. I am self-taught, adaptable and flexible to new environments and new technologies. Always seeking for delivering best results to the business.
  • University

  • B.S in Computer Science from USJT - 2020
  • MBA in Data Engineer from IGTI - 2021-2022
  • Living in

    São Paulo, Brazil

    Portfolio

    Here you can find my developed projects: applications, mobile, frameworks, awards , certifications.
    I hope you enjoy.

    DATASUS - Depression
    Developed Projects

    Read More
    Park Pay Go
    Final Paper

    Read More
    Bradesco
    Developed Projects

    Read More
    University
    Programming Marathon

    Read More
    Certificate
    conclusion

    Read More
    Awards
    Achievement

    Read More

    Experience

    Timeline

    Data Engineer Pl. Working on projects in the Big Data area, using current technologies PySpark, Apache Spark, Apache Kafka, Azure DataFactory, Databricks, Google Cloud Platform, Microsoft Azure at BlueShift Brasil

    Dez 2020 - o momento · Calculando...
    Data Engineer Pl. Working on projects in the Big Data area, using current technologies PySpark, Apache Spark, Apache Kafka, Azure DataFactory, Databricks, Google Cloud Platform, Microsoft Azure at BlueShift Brasil
    Developed projects
    Natura - Performance in projects with the big data team: Development and maintenance of data pipeline components, primarily for data ingestion and data quality. Technologies: Apache Airflow Pyspark Jenkins Hadoop (Hive, HDFS, Yarn) AWS (EMR, S3, Redshift, Glue) Oracle, MySQL, SQLServer Databricks Workflows Catalog Explorer Product Hive metastore Workflows using the Databricks environment to orchestrate ETL (Extract, Transform, Load) processes and data ingestion in Big Data environments. I utilize Databricks as the primary platform for building and executing data ingestion jobs, ensuring high availability and scalability. Additionally, I have experience in creating data products through the Catalog Explorer and Hive Metastore, enabling me to implement robust and efficient solutions for managing and cataloging metadata in distributed storage environments.

    Set 2022 - o momento · Calculando...
    Data Engineer Jr. Working on projects in the Big Data area, using current technologies PySpark, Apache Spark, Apache Kafka, Azure DataFactory, Databricks, Google Cloud Platform, Microsoft Azure at BlueShift Brasil
    Developed projects
    Natura - Performance in projects with the big data team: Development and maintenance of data pipeline components, primarily for data ingestion and data quality. Technologies: Apache Airflow Pyspark Jenkins Hadoop (Hive, HDFS, Yarn) AWS (EMR, S3, Redshift, Glue) Oracle, MySQL, SQLServer Databricks Workflows Catalog Explorer Product Hive metastore Workflows using the Databricks environment to orchestrate ETL (Extract, Transform, Load) processes and data ingestion in Big Data environments. I utilize Databricks as the primary platform for building and executing data ingestion jobs, ensuring high availability and scalability. Additionally, I have experience in creating data products through the Catalog Explorer and Hive Metastore, enabling me to implement robust and efficient solutions for managing and cataloging metadata in distributed storage environments.

    Mar 2021 - set 2022 · 1 ano 7 meses
    Data Engineer Jr. Working on projects in the Big Data area, using current technologies PySpark, Apache Spark, Apache Kafka, Azure DataFactory, Databricks, Google Cloud Platform, Microsoft Azure at BlueShift Brasil
    Developed projects
    Client Netshoes / Magalu - Performance in projects with the big data team: Development of ETLs in pyspark using Databricks and Dataproc, using data from BigQuery, sqldw, and google datalake (GFS). Creation of routines for sending encrypted data. Development of procedures and views in Oracle and SQL DW databases.

    Mar 2021
    Data Engineer Trainee - Big Data & Analytics I am currently specializing in the area of Big Data & Analytics at BlueShift Brasil
    Developed projects
    Depression: Integrated project that consists of collecting data on the theme of Depression, which is at the base of DATASUS, based on the data extraction process (ETL), through the Google Cloud Platform, the flow that was developed uses the Computer Engine to carry out the treatment of the base via Python and Cloud Storage (GCP), to make the storage of the treated base and, finally, BigQuery for SQL queries, finalizing the infrastructure of the base.

    Dec 2020 - mar 2021 4 meses
    Develop low platform workflow applications for information, flexible rates, creation of process dashboards internal areas, and segmentation of product information bases and services, using the Access tool to develop the SQL data, and VBA codes in Bradesco S.A
    2018 - 2020
    Fundação Florestal
    Information Technology Internship Technical support to users and customers, configuration of application databases: installation, configuration, services and monitoring (e-mail, Office, ERP and client software), Install and configure micros, Install and configure networks, Create spreadsheets electronic reports and miscellaneous reports, Support IT infrastructure, solving problems in systems and programs.
    Fundação para Conservação e Produção Florestal do Estado de São Paulo
    2017 - 2018

    Skills

    Always studying for Implementing the most current methodologies and techniques.

    Languages

    Javascript
    Angular
    HTML5
    Python
    MySQL
    CSS3
    VBA

    Technologies & Tools

    Amazon Web Services
    Google Cloud Plataform
    Microsoft Azure
    Apache Spark
    Apache Kafka
    Illustrator
    Windows
    VSCode
    MacOS
    Github
    Git

    Currently learning

    Cloud Computer
    Data Factory
    ETL in scale
    Big Data
    Hadoop
    PySpark
    Python
    NoSQL

    Blog

    My blog

    November 21, 2020 Github

    How to place a banner in your overview GitHub

    You can share information about yourself with the community on GitHub by creating a profile README. GitHub shows your profile README at the top of your profile page.

    Contact

    Do you have any suggestions, requests or do you want to talk about something, write me a message or find me on social networks.
    Thank you!