<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Etcd on Taubyte Blog</title><link>/blog/tags/etcd/</link><description>Recent content in Etcd on Taubyte Blog</description><image><title>Taubyte Blog</title><url>/blog/opengraph.jpg</url><link>/blog/opengraph.jpg</link></image><generator>Hugo -- 0.146.0</generator><language>en-us</language><lastBuildDate>Thu, 19 Feb 2026 16:00:00 +0000</lastBuildDate><atom:link href="/blog/tags/etcd/index.xml" rel="self" type="application/rss+xml"/><item><title>etcd: The Consensus Tax You're Probably Paying For Nothing</title><link>/blog/posts/etcd-performance-fresser/</link><pubDate>Thu, 19 Feb 2026 16:00:00 +0000</pubDate><guid>/blog/posts/etcd-performance-fresser/</guid><description>&lt;p>etcd sits at the heart of Kubernetes. Before your applications run, etcd is storing cluster state, coordinating elections, and replicating data. It consumes 2-8 GB RAM per node. It requires 3-5 nodes for high availability. That&amp;rsquo;s 6-40 GB RAM just for cluster coordination.&lt;/p>
&lt;p>Most teams don&amp;rsquo;t need distributed consensus. Most teams don&amp;rsquo;t need high availability at the cluster level. Most teams are running small clusters that would work fine with a single node and backups.&lt;/p></description></item><item><title>Why Raft Fails in Production and How Taubyte Raft Fixes It</title><link>/blog/posts/why-raft-fails-in-production-and-how-taubyte-raft-fixes-it/</link><pubDate>Thu, 29 Jan 2026 12:00:00 +0000</pubDate><guid>/blog/posts/why-raft-fails-in-production-and-how-taubyte-raft-fixes-it/</guid><description>Most Raft implementations look great in theory and fall apart in practice. The algorithm itself isn&amp;rsquo;t the problem—it&amp;rsquo;s everything around the algorithm that breaks in production: bootstrapping, discovery, leader routing, rejoin behavior, and what happens when nodes start out of order or the network is unreliable. Taubyte&amp;rsquo;s Raft wraps HashiCorp Raft and adapts it with libp2p transport, Taubyte discovery, and datastore-backed persistence. The goal isn&amp;rsquo;t to reinvent consensus—it&amp;rsquo;s to make consensus operable. Nodes can start in any order and converge to a working cluster without static seed lists or fragile bootstrap rituals. This article explores how Taubyte&amp;rsquo;s Raft addresses the operational challenges that make Kubernetes/etcd fragile and compares it to typical Raft libraries.</description></item></channel></rss>