<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Monitoring on Build. Run. Repeat.</title><link>https://buildrunrepeat.com/tags/monitoring/</link><description>Recent content in Monitoring on Build. Run. Repeat.</description><generator>Hugo</generator><language>en</language><lastBuildDate>Wed, 01 Jan 2025 09:00:00 -0400</lastBuildDate><atom:link href="https://buildrunrepeat.com/tags/monitoring/index.xml" rel="self" type="application/rss+xml"/><item><title>HashiCorp Consul Service Mesh on Kubernetes Series - Part 1 - Introduction and Setup</title><link>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-01-intro-and-setup/</link><pubDate>Wed, 01 Jan 2025 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-01-intro-and-setup/</guid><description>&lt;p&gt;Modern cloud-native architectures rely heavily on microservices, and Kubernetes has become the go-to platform for deploying, managing, and scaling these distributed applications. As the number of microservices grows, ensuring secure, reliable, and observable service-to-service communication becomes increasingly complex. This is where service mesh solutions, such as HashiCorp Consul, step in to provide a seamless approach to managing these challenges. In this blog post, we will delve into the integration of HashiCorp Consul Service Mesh with Kubernetes, exploring its architecture, features, and step-by-step deployment guide.&lt;/p&gt;</description></item><item><title>HashiCorp Consul Service Mesh on Kubernetes Series - Part 2 - Observability</title><link>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-02-observability/</link><pubDate>Wed, 01 Jan 2025 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-02-observability/</guid><description>&lt;p&gt;Modern service meshes require robust observability to ensure seamless operations, proactive troubleshooting, and performance optimization. In this section, we explore the observability features of HashiCorp Consul Service Mesh, including visualizing the service mesh, querying metrics, distributed tracing, and logging and auditing.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id="visualizing-the-service-mesh"&gt;Visualizing the Service Mesh&lt;/h2&gt;
&lt;p&gt;The Consul UI is used for visualizing the service mesh and its topology.&lt;/p&gt;
&lt;p&gt;Use the &lt;code&gt;watch&lt;/code&gt; command to send requests to the application continually. Make sure HTTP status code &lt;code&gt;200&lt;/code&gt; is returned in the output.&lt;/p&gt;</description></item><item><title>HashiCorp Consul Service Mesh on Kubernetes Series - Part 4 - Security</title><link>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-04-security/</link><pubDate>Wed, 01 Jan 2025 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/hashicorp-consul-k8s-service-mesh-series-04-security/</guid><description>&lt;p&gt;Security is a fundamental aspect of any service mesh, ensuring that all service-to-service communication is secure, controlled, and auditable. HashiCorp Consul provides robust security features, including mutual TLS (mTLS), access control, and rate limiting.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id="mtls"&gt;mTLS&lt;/h2&gt;
&lt;p&gt;In this section, we will demonstrate mTLS with Consul. Consul enables and strictly enforces mTLS by default. All traffic sent through the Consul Connect Service Mesh is encrypted.&lt;/p&gt;
&lt;p&gt;This section is slightly different from the Istio mTLS section because:&lt;/p&gt;</description></item><item><title>CAPV: Addressing Node Provisioning Issues Due to an Invalid State of ETCD</title><link>https://buildrunrepeat.com/posts/capv-addressing-node-provisioning-issues-due-to-invalid-state-of-etcd/</link><pubDate>Fri, 01 Dec 2023 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/capv-addressing-node-provisioning-issues-due-to-invalid-state-of-etcd/</guid><description>&lt;p&gt;I recently ran into a strange scenario on a Kubernetes cluster after a sudden and unexpected crash it had experienced due to an issue in the underlying vSphere environment. In this case, the cluster was a TKG cluster (in fact, it happened to be the TKG management cluster), however, the same situation could have occurred on any cluster managed by Cluster API Provider vSphere (CAPV).&lt;/p&gt;
&lt;p&gt;I have seen clusters unexpectedly crash many times before and most of the time, they successfully went back online when all nodes were up and running. In this case, however, some of the nodes could not boot properly, and Cluster API started attempting their reconciliation.&lt;/p&gt;</description></item><item><title>TKG 2.3: Fixing the Prometheus Data Source in the Grafana Package</title><link>https://buildrunrepeat.com/posts/tkg-2-3-fixing-the-prometheus-data-source-in-the-grafana-package/</link><pubDate>Fri, 01 Sep 2023 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/tkg-2-3-fixing-the-prometheus-data-source-in-the-grafana-package/</guid><description>&lt;p&gt;With the release of TKG 2.3, the Grafana package was finally updated from version 7.5.x to 9.5.1.
If you have deployed the new Grafana package (&lt;code&gt;9.5.1+vmware.2-tkg.1&lt;/code&gt;) or upgraded your existing one to this version, you may have run into error messages in your Grafana dashboards.&lt;/p&gt;
&lt;p&gt;For example, in the &lt;code&gt;TKG Kubernetes cluster monitoring&lt;/code&gt; default dashboard, you may have run into the &lt;code&gt;Failed to call resource&lt;/code&gt; error when opening the dashboard and noticed that a lot of the data is missing.&lt;/p&gt;</description></item><item><title>Tanzu Kubernetes Grid GPU Integration</title><link>https://buildrunrepeat.com/posts/tkg-gpu-integration/</link><pubDate>Wed, 01 Mar 2023 09:00:00 -0400</pubDate><guid>https://buildrunrepeat.com/posts/tkg-gpu-integration/</guid><description>&lt;p&gt;I recently had to demonstrate Tanzu Kubernetes Grid and its GPU integration capabilities.
Developing a good use case and assembling the demo required some preliminary research.&lt;/p&gt;
&lt;p&gt;During my research, I reached out to Jay Vyas, staff engineer at VMware, SIG Windows lead for Kubernetes, a Kubernetes legend, and an awesome guy in general. :) For those who don&amp;rsquo;t know Jay, he is also one of the authors of the fantastic book &lt;code&gt;Core Kubernetes&lt;/code&gt; (look it up!).&lt;/p&gt;</description></item></channel></rss>