Name: Guilherme de A. Gasque
Type: User
Company: Next Reason
Bio: Data Architect @ Next Reason | Mastering in Innovation Technology @ UNIFESP | Community Director @ Think Big Think Data | 🐍 Pyth♥n
Twitter: guigasque
Location: Campinas, SP - Brazil
Blog: https://www.linkedin.com/in/guigasque/
Guilherme de A. Gasque's Projects
Canonical knowledge base of the Foundation for Public Code vereniging
Exemple code for Arduino IDE (Arduino / ESP8266 / ESP32)
A topic-centric list of HQ open datasets. PR ☛☛☛
DEPRECATED; see https://github.com/boot2docker/boot2docker/pull/1408
The Data Engineering Cookbook
An Interface for the Brazilian Public Healthcare Data Repository (DATASUS) for the R Language
Data processing pipeline that extracts raw data, applies ETL processes to prepare it for analysis, stores it in the Delta format on Amazon S3, and utilizes PySpark within a Jupyter environment running in a Docker container managed by Jenkins
Google Cloud Client Library for Python
Coordenadas geográficas de fronteiras (em KML e JSON) de estados e municípios brasileiros
Complete data engineering pipeline running on Minikube Kubernetes, Argo CD, Spark, Trino, S3, Delta lake, Postgres+ Debezium CDC, MySQL,Airflow, Kafka Strimzi, Datahub, OpenMetadata,Zeppelin, Jupyter, JFrog Container Registry
Serverless Reference Architecture for Real-time Stream Processing
Model, visualizations, and animation of the Lorenz system
A registry of publicly available datasets on AWS
How to query MongoDB through SSH Tunnel with Python
It covers the creation of a PySpark image for processing data in S3, particularly in Delta format, as well as the intricacies of configuring PySpark within an existing Kubernetes environment running services on the python:3.9-slim-buster image.
An R package for reading data in the DBC (compressed DBF) format used by DATASUS.
A utility for mocking out the Python Requests library.
R package for (down)loading data from IBGE (Instituto Brasileiro de Geografia e Estatística)
Send data into Slack using this GitHub Action!
Operator for Apache Spark-on-Kubernetes for Stackable Data Platform
Kubernetes operator for managing the lifecycle of Apache Spark applications on Kubernetes.