Postgres in the time of monster hardware
https://www.enterprisedb.com/blog/postgres-time-monster-hardware
I don't know if you followed the release of the last generation of CPUs. AMD's latest Genoa CPU (AMD EPYC™ 9965) can run 768 threads. It has 192 cores per socket and 2 threads per core, with 2 sockets. Imagine adding 10 TB of RAM to such a beast! Of course, everyone will think of how useful it will be for virtualization. As a database person, I'd rather ask myself what Postgres could do with so many resources. I love simplicity in architecture. But I often meet customers with huge resource needs. With average hosts nowadays, the best answer for them is sometimes multi-parallel processing (MPP).
So, with this new hardware, can we stop using horizontal scalability? To understand the impact of running PostgreSQL on it, we must examine a few technical limits. The analysis will begin with NUMA (Non-Uniform Memory Access) architecture. Next, we will address I/O bandwidth limits. They are a big factor, no matter the CPU or memory. Next, we will look at how PostgreSQL behaves with many connections. This topic has historical limits that bring up key questions. Finally, we will test parallel queries. We will examine their scalability and effectiveness on systems with many CPU threads.
https://www.enterprisedb.com/blog/postgres-time-monster-hardware
Nping
https://github.com/hanshuaikang/Nping
Nping is a Ping tool developed in Rust. It supports concurrent Ping for multiple addresses, visual chart display, real-time data updates, and other features.
https://github.com/hanshuaikang/Nping
kvm
https://github.com/jetkvm/kvm
JetKVM is a high-performance, open-source KVM over IP (Keyboard, Video, Mouse) solution designed for efficient remote management of computers, servers, and workstations. Whether you're dealing with boot failures, installing a new operating system, adjusting BIOS settings, or simply taking control of a machine from afar, JetKVM provides the tools to get it done effectively.
https://github.com/jetkvm/kvm
tailpipe
https://github.com/turbot/tailpipe
Tailpipe is the lightweight, developer-friendly way to query logs.
Cloud logs, SQL insights. Collects logs from cloud, container and application sources. Query and analyze your data instantly with the power of SQL, right from your terminal.
Fast, local, and efficient. Runs locally, powered by DuckDB's in-memory analytics and Parquet's optimized storage.
An ecosystem of prebuilt intelligence. MITRE ATT&CK-aligned queries, prebuilt detections, benchmarks, and dashboards, all open source and community-driven.
Built to build with. Define detections as code, extend functionality with plugins and write custom SQL queries.
https://github.com/turbot/tailpipe
yaak
https://github.com/mountain-loop/yaak
Yaak is a desktop API client for interacting with REST, GraphQL, Server Sent Events (SSE), WebSocket, and gRPC APIs.
https://github.com/mountain-loop/yaak
It's a log eat log world!
https://obakeng.substack.com/p/its-a-log-eat-log-world
Let's discuss logging - unstructured, structured and canonical log lines - what they are and what value they bring to your production systems.
https://obakeng.substack.com/p/its-a-log-eat-log-world
Redis as a Primary Database for Complex Applications
https://faun.pub/redis-as-a-primary-database-for-complex-applications-501ced31f923
How Redis can be used as a primary database for complex applications that need to store data in multiple formats?
https://faun.pub/redis-as-a-primary-database-for-complex-applications-501ced31f923
Slicing Up—and Iterating on—SLOs
https://www.honeycomb.io/blog/slicing-up-and-iterating-on-slos
One of the main pieces of advice about Service Level Objectives (SLOs) is that they should focus on the user experience. Invariably, this leads to people further down the stack asking, “But how do I make my work fit the users?”—to which the answer is to redefine what we mean by “user.” In the end, a user is anyone who uses whatever it is you’re measuring.
https://www.honeycomb.io/blog/slicing-up-and-iterating-on-slos
Unlocking Kubernetes Observability with the OpenTelemetry Operator
https://www.dash0.com/blog/unlocking-kubernetes-observability-with-the-opentelemetry-operator
https://www.dash0.com/blog/unlocking-kubernetes-observability-with-the-opentelemetry-operator
terraform-backend-git
https://github.com/plumber-cd/terraform-backend-git
Terraform HTTP Backend implementation that uses Git repository as storage
https://github.com/plumber-cd/terraform-backend-git
tfbuddy
https://github.com/zapier/tfbuddy
TFBuddy allows Terraform Cloud users to get apply-before-merge workflows in their Pull Requests.
https://github.com/zapier/tfbuddy
Securing Continuous Delivery: Argo CD Threat Detection
https://medium.com/exness-blog/securing-continuous-delivery-argo-cd-threat-detection-6296f97c9197
https://medium.com/exness-blog/securing-continuous-delivery-argo-cd-threat-detection-6296f97c9197
Kubernetes: How kube-proxy and CNI Work Together
https://medium.com/@rifewang/kubernetes-how-kube-proxy-and-cni-work-together-1255d273f291
https://medium.com/@rifewang/kubernetes-how-kube-proxy-and-cni-work-together-1255d273f291
A practical guide to Kubernetes Gateway API
https://www.spectrocloud.com/blog/practical-guide-to-kubernetes-gateway-api
https://www.spectrocloud.com/blog/practical-guide-to-kubernetes-gateway-api
Ensuring Effective Helm Charts with Linting, Testing, and Diff Checks
https://dev.to/hkhelil/ensuring-effective-helm-charts-with-linting-testing-and-diff-checks-ni0
https://dev.to/hkhelil/ensuring-effective-helm-charts-with-linting-testing-and-diff-checks-ni0
Metal3
https://metal3.io
Metal3 (pronounced “metal cubed”) is an open-source project that provides a set of tools for managing bare-metal infrastructure using Kubernetes.
https://metal3.io
autotune
https://github.com/kruize/autotune
Kruize Autotune is an Autonomous Performance Tuning Tool for Kubernetes. Autotune accepts a user provided "slo" goal to optimize application performance. It uses Prometheus to identify "layers" of an application that it is monitoring and matches tunables from those layers to the user provided slo. It then runs experiments with the help of a hyperparameter optimization framework to arrive at the most optimal values for the identified set of tunables to get a better result for the user provided slo.
Autotune can take an arbitrarily large set of tunables and run experiments to continually optimize the user provided slo in incremental steps. For this reason, it does not necessarily have a "best" value for a set of tunables, only a "better" one than what is currently deployed.
https://github.com/kruize/autotune
kubeip
https://github.com/doitintl/kubeip
Kubernetes' nodes don't necessarily need their own public IP addresses to communicate. However, there are certain situations where it's beneficial for nodes in a node pool to have their own unique public IP addresses.
For instance, in gaming applications, a console might need to establish a direct connection with a cloud virtual machine to reduce the number of hops.
Similarly, if you have multiple agents running on Kubernetes that need a direct server connection, and the server needs to whitelist all agent IPs, having dedicated public IPs can be useful. These scenarios, among others, can be handled on a cloud-managed Kubernetes cluster using Node Public IP.
KubeIP is a utility that assigns a static public IP to each node it manages. The IP is allocated to the node's primary network interface, chosen from a pool of reserved static IPs using platform-supported filtering and ordering.
If there are no static public IPs left, KubeIP will hold on until one becomes available. When a node is removed, KubeIP releases the static public IP back into the pool of reserved static IPs.
https://github.com/doitintl/kubeip
The case of the vanishing CPU: A Linux kernel debugging story
https://clickhouse.com/blog/a-case-of-the-vanishing-cpu-a-linux-kernel-debugging-story
A mysterious CPU spike in ClickHouse Cloud on GCP led to months of debugging, revealing a deeper issue within the Linux kernel’s memory management. What started as random performance degradation turned into a deep dive into kernel internals, where engineer Sergei Trifonov uncovered a hidden livelock. His journey through eBPF tracing, perf analysis, and a reproducible test case ultimately led to a surprising fix - only for another kernel bug to surface right after. Curious, read on…
https://clickhouse.com/blog/a-case-of-the-vanishing-cpu-a-linux-kernel-debugging-story