Jordi

banner pi cluster hadoop hdfs

Introduction

Weeks ago I decided to start creating an experimental home size "Big data" system based on Apache Spark. The first step for it is to create a distributed filesystem where Apache Spark will read and write eveything.

HDFS is the Hadoop distributed filesystem which provides features like: fault detection and recovery, huge datasets, hardware at data, etc... despite it is a Hadoop ecosystem piece, it works nice as the data distributed filesytem for Apache Spark.


(...)
[Continue Reading]

banner-pi-cluster-kubespray.jpg

Introduction

Recently, I've been thinking about improving my Kubernetes (K8s) Raspberry Pi cluster to  something more resilient than just a one K8s master cluster which can fail easily and ruin all your work of months on it.
I installed multiple times on-premise single master K8s clusters based on Redhat ecosystem ( Centos/Fedora). The difficulty of doing this manually if it's your first time it's far from easy, and if you want to install something closer to an on premises HA K8s cluster manually it becomes a quite uncomfortable task.
So the challenge was to find a simple way to install a multi-master K8s cluster using my new 10 raspberry Pi 4 (4Gb Ram) project. 

(...)
[Continue Reading]