diff --git a/markdown/Build-your-customizable-voice-assistant-with-Platypush.md b/markdown/Build-your-customizable-voice-assistant-with-Platypush.md index a754c60..15c68e1 100644 --- a/markdown/Build-your-customizable-voice-assistant-with-Platypush.md +++ b/markdown/Build-your-customizable-voice-assistant-with-Platypush.md @@ -95,7 +95,7 @@ First things first: in order to get your assistant working you’ll need: I’ll also assume that you have already installed Platypush on your device — the instructions are provided on the [Github page](https://git.platypush.tech/platypush/platypush), on -the [wiki](https://git.platypush.tech/platypush/platypush/-/wikis/home#installation) and in +the [wiki](https://git.platypush.tech/platypush/platypush/wiki/Home#installation) and in my [previous article](https://blog.platypush.tech/article/Ultimate-self-hosted-automation-with-Platypush). Follow these steps to get the assistant running: diff --git a/markdown/Detect-people-with-a-RaspberryPi-a-thermal-camera-Platypush-and-a-pinch-of-machine-learning.md b/markdown/Detect-people-with-a-RaspberryPi-a-thermal-camera-Platypush-and-a-pinch-of-machine-learning.md index 7c0cef2..d81a3a2 100644 --- a/markdown/Detect-people-with-a-RaspberryPi-a-thermal-camera-Platypush-and-a-pinch-of-machine-learning.md +++ b/markdown/Detect-people-with-a-RaspberryPi-a-thermal-camera-Platypush-and-a-pinch-of-machine-learning.md @@ -95,8 +95,8 @@ camera module. This post assumes that you have already installed and configured Platypush on your system. If not, head to my post on [getting started with Platypush](https://blog.platypush.tech/article/Ultimate-self-hosted-automation-with-Platypush), the [readthedocs page](https://docs.platypush.tech/en/latest/), the -[Gitlab page](https://git.platypush.tech/platypush/platypush) or -[the wiki](https://git.platypush.tech/platypush/platypush/-/wikis/home). +[repository page](https://git.platypush.tech/platypush/platypush) or +[the wiki](https://git.platypush.tech/platypush/platypush/wiki). Install also the Python dependencies for the HTTP server, the MLX90640 plugin and Tensorflow: diff --git a/markdown/One-browser-extension-to-rule-them-all.md b/markdown/One-browser-extension-to-rule-them-all.md index 53b3ca9..308e3b8 100644 --- a/markdown/One-browser-extension-to-rule-them-all.md +++ b/markdown/One-browser-extension-to-rule-them-all.md @@ -475,7 +475,7 @@ no cloud!), or loaded from a URL. ## Work in Progress The extension is still under development, and I’m open to suggestions, tickets, and pull requests on the -[Gitlab page](https://git.platypush.tech/platypush/platypush-webext). Two features, in particular, are next on my +[repository page](https://git.platypush.tech/platypush/platypush-webext). Two features, in particular, are next on my roadmap: ### Integration with the Platypush WebSocket protocol diff --git a/markdown/Ultimate-self-hosted-automation-with-Platypush.md b/markdown/Ultimate-self-hosted-automation-with-Platypush.md index f75696b..be0f707 100644 --- a/markdown/Ultimate-self-hosted-automation-with-Platypush.md +++ b/markdown/Ultimate-self-hosted-automation-with-Platypush.md @@ -68,7 +68,7 @@ a few sensors on a small Raspberry Zero is guaranteed to take not more than 5-10 The flexibility of Platypush comes however a slightly steeper learning curve, but it rewards the user with much more room for customization. You are expected to install it via [pip](https://pypi.org/project/platypush/) or -the [Gitlab](https://git.platypush.tech/platypush/platypush) repo, install the dependencies based on the plugins you +the [git repo](https://git.platypush.tech/platypush/platypush), install the dependencies based on the plugins you want (although managing per-plugin dependencies is quite easy via `pip`), and manually create or edit a configuration file. But it provides much, much more flexibility. It can listen for messages on MQTT, HTTP (but you don’t have to run the webserver if you don’t want to), websocket, TCP socket, Redis, Kafka, Pushbullet — you name it, it has probably got