| ▲ | valenterry 15 hours ago |
| So, let's say you want to deploy server instances. Let's keep it simple and say you want to have 2 instances running. You want to have zero-downtime-deployment. And you want to have these 2 instances be able to access configuration (that contains secrets). You want load balancing, with the option to integrate an external load balancer. And, last, you want to be able to run this setup both locally and also on at least 2 cloud providers. (EDIT: I meant to be able to run it on 2 cloud providers. Meaning, one at a time, not both at the same time. The idea is that it's easy to migrate if necessary) This is certainly a small subset of what kubernetes offers, but I'm curious, what would be your goto-solution for those requirements? |
|
| ▲ | bruce511 15 hours ago | parent | next [-] |
| That's an interesting set of requirements though. If that is indeed your set of requirements then perhaps Kubernetes is a good choice. But the set seems somewhat arbitrary. Can you reduce it further? What if you don't require 2 cloud providers? What if you don't need zero-downtime? Indeed given that you have 4 machines (2 instances, x 2 providers) could a human manage this? Is Kubernetes overkill? I ask this merely to wonder. Naturally if you are rolling out hundreds of machines you should, and no doubt by then you have significant revenue (and thus able to pay for dedicated staff) , but where is the cross-over? Because to be honest most startups don't have enough traction to need 2 servers, never mind 4, never mind 100. I get the aspiration to be large. I get the need to spend that VC cash. But I wonder if Devops is often just premature and that focus would be better spent getting paying customers. |
| |
| ▲ | valenterry 14 hours ago | parent [-] | | > Can you reduce it further? What if you don't require 2 cloud providers? What if you don't need zero-downtime? I think the "2 cloud providers" criteria is maybe negotiable. Also, maybe there was a misunderstanding: I didn't mean to say I want to run it on two cloud providers. But rather that I run it on one of them but I could easily migrate to the other one if necessary. The zero-downtime one isn't. It's not necessarily so much about actually having zero-downtime. It's about that I don't want to think about it. Anything besides zero-downtime actually adds additional complexity to the development process. It has nothing to do with trying to be large actually. | | |
| ▲ | AznHisoka 14 hours ago | parent [-] | | I disagree with that last part. By default, having a few seconds downtime is not complex. The easiest thing you could do to a server is restart it. Its literally just a restart! | | |
| ▲ | valenterry 13 hours ago | parent | next [-] | | It's not. Imagine there is a bug that stops the app from starting. It could be anything, from a configuration error (e.g. against the database) to a problem with warmup (if necessary) or any kind of other bug like an exception that only triggers in production for whatever reasons. EDIT: and worse, it could be something that just started and would even happen when trying to deploy the old version of the code. Imagine a database configuration change that allows the old connections to stay open until they are closed but prevents new connections from being created. In that case, even an automatic roll back to the previous code version would not resolve the downtime. This is not theory, I had those cases quite a few times in my career. | |
| ▲ | globular-toast 11 hours ago | parent | prev [-] | | I managed a few production services like this and it added a lot of overhead to my work. On the one hand I'd get developers asking me why their stuff hasn't been deployed yet. But then I'd also have to think carefully about when to deploy and actually watch it to make sure it came back up again. I would often miss deployment windows because I was doing something else (my real job). I'm sure there are many solutions but K8s gives us both fully declarative infrastructure configs and zero downtime deployment out of the box (well, assuming you set appropriate readiness probes etc) So now I (a developer) don't have to worry about server restarts or anything for normal day to day work. We don't have a dedicated DevOps/platforms/SRE team or whatnot. Now if something needs attention, whatever it is, I put my k8s hat on and look at it. Previously it was like "hmm... how does this service deployment work again..?" |
|
|
|
|
| ▲ | osigurdson 15 hours ago | parent | prev | next [-] |
| "Imagine you are in a rubber raft, you are surrounded by sharks, and the raft just sprung a massive leak - what do you do?". The answer, of course, is to stop imagining. Most people on the "just use bash scripts and duct tape" side of things assume that you really don't need these features, that your customers are ok with downtime and generally that the project that you are working on is just your personal cat photo catalog anyway and don't need such features. So, stop pretending that you need anything at all and get a job at the local grocery store. The bottom line is there are use cases, that involve real customers, with real money that do need to scale, do need uptime guarantees, do require diverse deployment environments, etc. |
| |
| ▲ | QuiDortDine 14 hours ago | parent | next [-] | | Yep. I'm one of 2 Devops at an R&D company with about 100 employees. They need these services for development, if an important service goes down you can multiply that downtime by 100, turning hours into man-days and days into man-months. K8 is simply the easiest way to reduce the risk of having to plead for your job. I guess most businesses are smaller than this, but at what size do you start to need reliability for your internal services? | |
| ▲ | ozim 11 hours ago | parent | prev [-] | | You know that you can scale servers just as well, you can use good practices with scripts and deployments in bash and having them documented and in version control. Equating bash scripts and running servers to duct taping and poor engineering vs k8s yaml being „proper engineering„ is well wrong. |
|
|
| ▲ | caseyohara 15 hours ago | parent | prev | next [-] |
| I think you are proving the point; there are very, very few applications that need to run on two cloud providers. If you do, sure, use Kubernetes if that makes your job easier. For the other 99% of applications, it’s overkill. Apart from that requirement, all of this is very doable with EC2 instances behind an ALB, each running nginx as a reverse proxy to an application server with hot restarting (e.g. Puma) launched with a systemd unit. |
| |
| ▲ | osigurdson 15 hours ago | parent | next [-] | | To me that sounds harder than just using EKS. Also, other people are more likely to understand how it works, can run it in other environments (e.g. locally), etc. | |
| ▲ | valenterry 14 hours ago | parent | prev | next [-] | | Sorry, that was a misunderstanding. I meant that I want to be able to run it on two cloud providers, but one at a time is fine. It just means that it would be easy to migrate/switch over if necessary. | |
| ▲ | globular-toast 11 hours ago | parent | prev [-] | | Hmm, let's see, so you've got to know: EC2, ALB, Nginx, Puma, Systemd, then presumably something like Terraform and Ansible to deploy those configs, or write a custom set of bash scripts. And all of that and you're tied to one cloud provider. Or, instead of reinventing the same wheels for Nth time, I could just use a set of abstractions that work for 99% of network services out there, on any cloud or bare metal. That set of abstractions is k8s. | | |
|
|
| ▲ | tootubular 15 hours ago | parent | prev | next [-] |
| My personal goto-solution for those requirements -- well 1 cloud provider, I'll follow up on that in a second -- would be using ECS or an equivalent service. I see the OP was a critic of Docker as well, but for me, ECS hits a sweet spot. I know the compute is at a premium, but at least in my use-cases, it's so far been a sensible trade. About the 2 cloud providers bit. Is that a common thing? I get wanting migrate away from one for another, but having a need for running on more than 1 cloud simultaneously just seems alien to me. |
| |
| ▲ | mkesper 8 hours ago | parent | next [-] | | Last time I checked ECS was even more expensive than using Lambda but without the ability of fast starting your container, so I really don't get the niche it fits into, compared to Lambda on one side and self-hosting docker on minimal EC2 instances on the other side. | | |
| ▲ | tootubular 6 hours ago | parent [-] | | I may need to look at Lambda closer! At least way back, I thought it was a no-go since the main runtime I work with is Ruby. As for minimal EC2 instances, definitely, I do that for environments where it makes sense and that's the case fairly often. |
| |
| ▲ | valenterry 14 hours ago | parent | prev [-] | | Actually, I totally agree. ECS (in combination with secret manager) is basically fulfilling all needs, except being not so easy to reproduce/simulate locally and of course with the vendor lock-in. |
|
|
| ▲ | shrubble 15 hours ago | parent | prev | next [-] |
| Do you know of actual (not hypothetical) cases, where you could "flip a switch" and run the exact same Kubernetes setups on 2 different cloud providers? |
| |
| ▲ | InvaderFizz 14 hours ago | parent | next [-] | | I run clusters on OKE, EKS, and GKE. Code overlap is like 99% with the only real differences all around ingress load balancers. Kubernetes is what has provided us the abstraction layer to do multicloud in our SaaS. Once you are outside the k8s control plane, it is wildly different, but inside is very consistent. | |
| ▲ | threeseed 15 hours ago | parent | prev | next [-] | | Yes. I've worked on a number of very large banking and telco Kubernetes platforms. All used multi-cloud and it was about 95% common code with the other 5% being driver style components for underlying storage, networking, IAM etc. Also using Kind/k3d for local development. | |
| ▲ | devops99 15 hours ago | parent | prev | next [-] | | Both EKS (Amazon) and GKE (Google Cloud) run Cilium for the networking part of their managed Kubernetes offerings. That's the only real "hard part". From the users' point of view, the S3 buckets, the network-attached block devices, and compute (CRIO container runtime) are all the same. You are using some other cloud provider or want uniformity there's https://Talos.dev | |
| ▲ | brodo 6 hours ago | parent | prev | next [-] | | If you are located in germany and run critial IT infrastructure (banks, insurance companies, energy companies) you have to be able to deal with a cloud provider completely going down in 24 houres. Not everyone who has to can really do it, but the big players can. | |
| ▲ | hi_hi 15 hours ago | parent | prev [-] | | Yes, but it would involve first setting up a server instance and then installing k3s :-) | | |
| ▲ | valenterry 13 hours ago | parent [-] | | I actually also think that k3s probably comes closest to that. But I have never used it, and ultimately it also uses k8s. |
|
|
|
| ▲ | kccqzy 14 hours ago | parent | prev | next [-] |
| I've worked at tiny startups before. Tiny startups don't need zero-downtime-deployment. They don't have enough traffic to need load balancing. Especially when you are running locally, you don't need any of these. |
| |
| ▲ | anon7000 14 hours ago | parent | next [-] | | Tiny startups can’t afford to loose customers because they can’t scale though, right? Who is going to invest in a company that isn’t building for scale? Tiny startups are rarely trying to build projects for small customer bases (eg little scaling required.) They’re trying to be the next unicorn. So they should probably make sure they can easily scale away from tossing everything on the same server | | |
| ▲ | lmm 13 hours ago | parent | next [-] | | > Tiny startups can’t afford to loose customers because they can’t scale though, right? Who is going to invest in a company that isn’t building for scale? Having too many (or too big) customers to handle is a nice problem to have, and one you can generally solve when you get there. There are a handful of giant customers that would want you to be giant from day 1, but those customers are very difficult to land and probably not worth the effort. | |
| ▲ | jdlshore 13 hours ago | parent | prev [-] | | Startups need product-market fit before they need scale. It’s incredibly hard to come by and most won’t get it. Their number one priority should be to run as many customer acquisition experiments as possible for as little as possible. Every hour they spend on scale before they need it is an hour less of runway. | | |
| ▲ | lkjdsklf 12 hours ago | parent [-] | | while true, zero downtime deployments is... trivial... even for a tiny startup.. So you might as well do it. | | |
|
| |
| ▲ | p_l 4 hours ago | parent | prev [-] | | Tiny startups don't have money to spend on too much PaaS or too many VMs or faff around with custom scripts for all sorts of work. Admittedly, if you don't know k8s, it might be non-starter... but if you some knowledge, k3s plus cheap server is a wonderful combo |
|
|
| ▲ | whatever1 15 hours ago | parent | prev | next [-] |
| Why does a startup need zero-downtime-deployment? Who cares if your site is down for 5 seconds? (This is how long it takes to restart my Django instance after updates). |
| |
| ▲ | valenterry 14 hours ago | parent | next [-] | | Because it increases development speed. It's maybe okay to be down for 5 seconds. But if I screw up, I might be down until I fix it. With zero-downtime deployment, if I screw up, then the old instances are still running and I can take my time to fix it. | |
| ▲ | everfrustrated 8 hours ago | parent | prev [-] | | If you're doing CD where every push is an automated deploy a small company might easily have a hundred deploys a day. So you need seamless deployments. | | |
| ▲ | xdennis an hour ago | parent [-] | | I think it's a bit of an exaggeration to say a "small" company easily does 100 deployments a day. |
|
|
|
| ▲ | lkjdsklf 14 hours ago | parent | prev | next [-] |
| We’ve been deploying software like this for a long ass time before kubernetes. There’s shitloads of solutions. It’s like minutes of clicking in a ui of any cloud provider to do any of that. So doing it multiple times is a non issue. Or automate it with like 30 lines of bash. Or chef. Or puppet. Or salt. Or ansible. Or terraform. Or or or or or. Kubernetes brings in a lot of nonsense that isn’t worth the tradeoff for most software. If you feel it makes your life better, then great! But there’s way simpler solutions that work for most things |
| |
| ▲ | valenterry 13 hours ago | parent [-] | | I'm actually not using kubernetes because I find it too complex. But I'm looking for a solution for that problem and I haven't found one, so I was wondering what OP uses. Sorry, but I don't want to "click in a UI". And it is certainly not something you can just automate with 30 lines of bash. If you can, please elaborate. | | |
| ▲ | lkjdsklf 12 hours ago | parent [-] | | > And it is certainly not something you can just automate with 30 lines of bash. If you can, please elaborate. Maybe not literally 30.. I didn't bother actually writing it. Also bash was just a single example. It's way less terraform code to do the same thing. You just need an ELB backed by an autoscaling group. That's not all that much to setup. That gets you the two loadbalanced servers and zero downtime deploys. When you want to deploy, you just create a new scaling group and launch configuration and attach to the ELB and ramp down the old one.. Easy peasy. For the secrets, you need at least KMS and maybe secret manager if you're feeling fancy.. That's not much to setup. I know for sure AWS and azure provide nice CLIs that would let you do this in not that many commands. or just use terraform Personally if I really cared about multi cloud support, I'd go terraform (or whatever it's called now). | | |
| ▲ | valenterry 12 hours ago | parent [-] | | > You just need an ELB backed by an autoscaling group Sure, and then you can neither 1.) test your setup locally nor 2.) easily move to another cloud provider. So that doesn't really fit what I asked. If they answer is "there is nothing, just accept the vendor lock-in" then fine, but please don't reply with "30 lines of bash" and make me have expectations. :-( |
|
|
|
|
| ▲ | rozap 13 hours ago | parent | prev | next [-] |
| A script that installs some dependencies on an Ubuntu vm. A script that rsyncs the build artifact to the machine. The script can drain connections and restart the service using the new build, then onto the next VM. The cloud load balancer points at those VMs and has a health check. It's very simple. Nothing fancy. Our small company uses this setup. We migrated from GCP to AWS when our free GCP credits from YC ran out and then we used our free AWS credits. That migration took me about a day of rejiggering scripts and another of stumbling around in the horrible AWS UI and API. Still seems far, far easier than paying the kubernetes tax. |
| |
| ▲ | valenterry 13 hours ago | parent [-] | | I guess the cloud load balancer is the most custom part. Do you use the alb from aws? |
|
|
| ▲ | wordofx 8 hours ago | parent | prev | next [-] |
| 0 downtime. Jesus Christ. Nginx and HAProxy solved this shit decades ago. You can drop out a server or group. Deploy it. Add it back in. With a single telnet command. You don’t need junk containers to solve things like “0 downtime deployments”. That was a solved problem. |
| |
| ▲ | valenterry 6 hours ago | parent [-] | | Calm down my friend! You are not wrong, but that only covers a part of what I was asking. How about the rest? How do you actually bring your services to production? I'm curious. And, PS, I don't use k8s. Just saying. |
|
|
| ▲ | amluto 15 hours ago | parent | prev | next [-] |
| For something this simple, multi-cloud seems almost irrelevant to the complexity. If I’m understanding your requirements right, a deployment consists of two instances and a load balancer (which could be another instance or something cloud-soecific). Does this really need to have fancy orchestration to launch everything? It could be done by literally clicking the UI to create the instances on a cloud and by literally running three programs to deploy locally. |
|
| ▲ | CharlieDigital 15 hours ago | parent | prev | next [-] |
| Serverless containers. Effectively using Google and Azure managed K8s. (Full GKE > GKE Autopilot > Google Cloud Run). The same containers will run locally, in Azure, or AWS. It's fantastic for projects but and small. The free monthly grant makes it perfect for weekend projects. |
|
| ▲ | gizzlon 6 hours ago | parent | prev [-] |
| Cloud Run. Did you read the article? Migrating to another cloud should be quite easy. There are many PaaS solutions.
The hard parts will be things like migrating the data, make sure there's no downtime AND no drift/diff in the underlying data when some clients write to Cloud-A and some write to CLoud-B, etc. But k8 do not fix these problems, so.. |
| |
| ▲ | htgb 5 hours ago | parent [-] | | Came here to say the same thing: PaaS. Intriguing that none of the other 12 sibling comments mention this… each in their bubble I guess (including me). We use Azure App Service at my day job and it just works. Not multi-cloud obviously, but the other stuff: zero downtime deploys, scale-out with load balancing… and not having to handle OS updates etc. And containers are optional, you can just drop your binaries and it runs. |
|