Kubernetes configuration linting tools
https://itnext.io/kubernetes-configuration-linting-tools-699ddeedaeec
https://itnext.io/kubernetes-configuration-linting-tools-699ddeedaeec
Git Happens: How Argo CD took over our deployments
https://mirakl.tech/git-happens-how-argo-cd-took-over-our-deployments-e214343e1532
https://mirakl.tech/git-happens-how-argo-cd-took-over-our-deployments-e214343e1532
Patroni Backups: When pgBackRest and ArgoCD Have Your Back (Literally)
https://medium.com/@yatzikziv/patroni-backups-when-pgbackrest-and-argocd-have-your-back-literally-091afa98be50
https://medium.com/@yatzikziv/patroni-backups-when-pgbackrest-and-argocd-have-your-back-literally-091afa98be50
(Yet) Another Take on Integrating Terraform with Argo CD
https://akuity.io/blog/yet-another-take-on-integrating-terraform-with-argo-cd
https://akuity.io/blog/yet-another-take-on-integrating-terraform-with-argo-cd
DBaaS in 2024: Which PostgreSQL operator for Kubernetes to select for your platform? Part 4
https://medium.com/@davidpech_39825/dbaas-in-2024-which-kubernetes-postgresql-operator-part-4-crunchys-pgo-9225d518c71d
https://medium.com/@davidpech_39825/dbaas-in-2024-which-kubernetes-postgresql-operator-part-4-crunchys-pgo-9225d518c71d
300,000+ Prometheus Servers and Exporters Exposed to DoS Attacks
https://www.aquasec.com/blog/300000-prometheus-servers-and-exporters-exposed-to-dos-attacks
https://www.aquasec.com/blog/300000-prometheus-servers-and-exporters-exposed-to-dos-attacks
Connecting Kubernetes K3s cluster to external router using BGP with MetalLB and Nginx Ingress
https://medium.com/@nikoolayy1/connecting-kubernetes-k3s-cluster-to-external-router-using-bgp-with-metallb-bgp-nginx-as-ingress-9bb767dcecd2
https://medium.com/@nikoolayy1/connecting-kubernetes-k3s-cluster-to-external-router-using-bgp-with-metallb-bgp-nginx-as-ingress-9bb767dcecd2
silver-surfer
https://github.com/devtron-labs/silver-surfer
Api-Version Compatibility Checker & Provides Migration Path for K8s Objects
https://github.com/devtron-labs/silver-surfer
kubectl-klock
https://github.com/applejag/kubectl-klock
A kubectl plugin to render the kubectl get pods --watch output in a much more readable fashion.
Think of it as running watch kubectl get pods, but instead of polling, it uses the regular watch feature to stream updates as soon as they occur.
https://github.com/applejag/kubectl-klock
1
kubepfm
https://github.com/flowerinthenight/kubepfm
kubepfm is a simple wrapper to the kubectl port-forward command for multiple pods/deployments/services. It can start multiple kubectl port-forward processes based on the number of input targets. Terminating the tool (Ctrl-C) will also terminate all running kubectl sub-processes.
https://github.com/flowerinthenight/kubepfm
oomd
https://github.com/facebookincubator/oomd
oomd is userspace Out-Of-Memory (OOM) killer for linux systems.
https://github.com/facebookincubator/oomd
cloud-snitch
https://github.com/ccbrown/cloud-snitch
Map visualization and firewall for AWS activity, inspired by Little Snitch for macOS.
https://github.com/ccbrown/cloud-snitch
arkflow
https://github.com/arkflow-rs/arkflow
High-performance Rust stream processing engine, providing powerful data stream processing capabilities, supporting multiple input/output sources and processors.
https://github.com/arkflow-rs/arkflow
brush
https://github.com/reubeno/brush
brush (Bo(u)rn(e) RUsty SHell) is a POSIX- and bash-compatible shell, implemented in Rust. It's built and tested on Linux and macOS, with experimental support on Windows. (Its Linux build is fully supported running on Windows via WSL.)
https://github.com/reubeno/brush
outpost
https://github.com/hookdeck/outpost
Outpost is a self-hosted and open-source infrastructure that enables event producers to add outbound webhooks and Event Destinations to their platform with support for destination types such as Webhooks, Hookdeck Event Gateway, Amazon EventBridge, AWS SQS, AWS SNS, GCP Pub/Sub, RabbitMQ, and Kafka.
https://github.com/hookdeck/outpost
tilt
https://github.com/tilt-dev/tilt
Define your dev environment as code. For microservice apps on Kubernetes.
https://github.com/tilt-dev/tilt
Anomaly Detection in Time Series Using Statistical Analysis
https://medium.com/booking-com-development/anomaly-detection-in-time-series-using-statistical-analysis-cc587b21d008
Setting up alerts for metrics isn’t always straightforward. In some cases, a simple threshold works just fine — for example, monitoring disk space on a device. You can just set an alert at 10% remaining, and you’re covered. The same goes for tracking available memory on a server.
But what if we need to monitor something like user behavior on a website? Imagine running a web store where you sell products. One approach might be to set a minimum threshold for daily sales and check it once a day. But what if something goes wrong, and you need to catch the issue much sooner — within hours or even minutes? In that case, a static threshold won’t cut it because user activity fluctuates throughout the day. This is where anomaly detection comes in.
https://medium.com/booking-com-development/anomaly-detection-in-time-series-using-statistical-analysis-cc587b21d008
Incident SEV scales are a waste of time
https://blog.danslimmon.com/2025/01/29/incident-sev-scales-are-a-waste-of-www.tgoop.com/
Ask an engineering leader about their incident response protocol and they’ll tell you about their severity scale. “The first thing we do is we assign a severity to the incident,” they’ll say, “so the right people will get notified.”
And this is sensible. In order to figure out whom to get involved, decision makers need to know how bad the problem is. If the problem is trivial, a small response will do, and most people can get on with their day. If it’s severe, it’s all hands on deck.
Severity correlates (or at least, it’s easy to imagine it correlating) to financial impact. This makes a SEV scale appealing to management: it takes production incidents, which are so complex as to defy tidy categorization on any dimension, and helps make them legible.
A typical SEV scale looks like this:
- SEV-3: Impact limited to internal systems.
- SEV-2: Non-customer-facing problem in production.
- SEV-1: Service degradation with limited impact in production.
- SEV-0: Widespread production outage. All hands on deck!
But when you’re organizing an incident response, is severity really what matters?
https://blog.danslimmon.com/2025/01/29/incident-sev-scales-are-a-waste-of-www.tgoop.com/