Reddit DevOps – Telegram
Buildstash - a platform for managing binaries and releases across apps/games/embedded

For a bit over a year now, I've been building a tool for teams to manage their software binaries and releases.

Obviously tools like Artifactory exist - but coming from an apps/games background I'd found the vast majority of teams didn't use any dedicated tool for managing binaries. Finding what's out there too complex / expensive / missing features around managing releases and deployment for projects not being deployed to a package manager.

A lot of Google Drive, SharePoint, and Slack dumping grounds - with context lost, and not really suited to keeping track of past builds, distribution, etc etc.

The idea and hope for Buildstash is to bring binary and release management to teams currently without a dedicated tool for it, making it so accessible even for small teams that it becomes as much a no-brainer as having source control or CI.

So, focusing on the features devs across app/games/embedded need for managing their builds and releases. Whether around collaboration (linking builds to related issues etc), integrated beta distribution, sharing build streams and releases on their website, and rolling out to distribution platforms like the App Store / Google Play/ Steam etc.

Here's a product demo video - https://youtu.be/t4Fr6M_vIIc

our landing - https://buildstash.com

and GitHub with various integrations - https://github.com/buildstash/

We're still at a really early stage but super proud of what we've built so far! I'd really love your feedback / experiences with this problem / thoughts on what we should build next? :)

https://redd.it/1nr5xlv
@r_devops
Who writes your technical documentation?

Which team is in charge of updating technical docs? Like all these chatbots in the docs section. Which department manages and updates them? DevRel? Product? Who does it for your org?

https://redd.it/1nr8pns
@r_devops
recently started job hunting, is my resume is good enough please can you gimme review on this.e

https://ibb.co/gL59j1Tr

https://ibb.co/KZyk3ZK

https://redd.it/1nr9mg1
@r_devops
💩1
Terraform Development in large teams

So we've had a consultancy waste investors money, I mean, understand the business, to, presumably, suggest job cuts.

Anyway, we're a small team of 3 and we have enough different things to get on with that it's very rare that we have two people working on the same project (terraform root module) at the same time AND become an issue with applies in dev.

If somebody needs to apply something, we just post in Teams that weirdness will happen in your plans and please don't apply until further notice.

Furthermore, we have a sandbox subnoscription for precisely these types of scenarios, namely apply something that we're not sure about and need to apply it first.

I'd say that we run into a scenario where somebody needs to apply to dev as part of their development about 1 a month. Most of the stuff tends to be routine, e.g. add a microservice number 28, we don't need to apply before merging to test that it will do the same that the other 27.

I explained this to the consultant and he went on about how this was a terrible way of working and he was surprised that we didn't run into issues more often. When I pointed out that I take reasonable good care to avoid this by ordering tickets he just said that this was just an accident waiting to happen and that we'd been very lucky.

I asked him how it was done in big teams and he said that you apply in dev and people then merge that feature branch into their feature branch to bring in those changes, he might've said cherry pick to be fair.

I asked him what happened if the original thing wasn't quite right, he said that you fix it, apply it and then everybody else incorporates the changes again.

To me this seems horrendously inefficient and requiring massive amounts of back-channel communication, which as the team increases in size is just going to create huge problems.

While I have worked at big teams (up to 10 engineers) we hardly ever had more than 2 people on the same thing so it's never been an issue

Just wonder how people do it in big teams.

https://redd.it/1nrb9cj
@r_devops
Looking for job 4+ Yoe

Hey all ,
I am looking out for new opportunities. Have 4.2 years of experience in devops / SRE.

Have skills in AWS, Splunk , Snyk, Gitlab CI/CD, Jenkins , Kubernetes, Terraform and python etc.

Also keen to learn and implement new tech tools.

I am from India . So let me know if you have something for me. Thanks

https://redd.it/1nrcne1
@r_devops
DevOps Hackathon by TRMNL (e-ink dashboard)

Hello DevOps OPs, commenters, and lurkers (i'm usually in the latter two). Starting today and going through Sunday, October 5, TRMNL is asking the question, "What would DevOps enthusiasts make for a TRMNL device?"

To answer that question, we're giving away $40 discount codes to all qualified entrants, and also TRMNL devices to 3 winners.

All the specs: https://usetrmnl.com/blog/hackathon-devops

If you're curious about what people have created (not just DevOps), check out the integrations and recipes.

https://redd.it/1nrgji3
@r_devops
Anyone have issues with AWS quota limits being inaccurate?

We're up to 140 vcpus in our account quota but we will run \~72 vcpus in fargate across scheduled one-off jobs but we get jobs rejected due to capacity constraints even when at the time we don't have instances active in our account.

I assume they either have a sliding window they use for quota accounting and we're just overwhelming it and need some sort of cool down which we've enacted by throttling to 1/3rd of our quota as the active queue concurrency.

Edit to add: Error is "Failed to run ECS task: You've reached the limit on the number of vCPUs you can run concurrently"

Anyone else seen this or happen to know any specifics on how the quotas are applied (e.g. per 60 second windows)?

https://redd.it/1nrhs8t
@r_devops
Impressions on my platform/devops resume

hi guys, I recently went back to school for my masters and am applying for internships, got a few OAs but they never convert to any interviews, let alone an offer and I won't count the rejections.

I know the market is bad atm, but I want to work on the things that are in my control and make the best out of my situation.

my resume on drive

https://redd.it/1nrjjrs
@r_devops
Looking for advice on scaling SEC data app (10 rps limit)

I’ve built a financial app that pulls company financials from the SEC—nearly verbatim (a few tags can be missing)—covering the XBRL era (2009/2010 to present). I’m launching a site to show detailed quarterly and annual statements.

Constraint: The SEC allows ~10 requests/second per IP, so I’m worried I can only support a few hundred concurrent users if I fetch on demand.

Goal: Scale beyond that without blasting the SEC and without storing/downloading the entire corpus.

What’s the best approach to:
• stay under ~10 rps to the SEC,
• keep storage minimal, and
• still serve fast, detailed statements to lots of users?

Any proven patterns (caching, precomputed aggregates, CDN, etc.) you’d recommend?

https://redd.it/1nrpunp
@r_devops
What in-house luxury dev tooling have you built?

At a previous job we had in house IDE extensions that checked if you were making backward incompatible changes that would break consumers in by checking against a service which held a graph of all the method usages between projects.

At current job I've made an integration that shows ci and argocd sync status badges inside our git browser.

These seem like to much effort to reinvent at next job but are were nice to have. Does your company have any cool or quirky custom tooling?

I am not secretly selling a product btw.


https://redd.it/1nrqu98
@r_devops
Zero downtime deployments with database migrations

I am looking for a solution where I can deploy my backend api changes and database migrations with 0 downtime.

I deploy my backend on azure container apps and use Azure Sql:
- I use container apps Multi revision mode to use blue green deployments. I already test green revisions to see if they are healthy or not.
- I create ef core migrations (idempotent)

The easiest solution I can think of (with the tools I currently use) is to block developers from adding migrations that have both Additions and Deletions.

I am wondering, how are you doing this?

https://redd.it/1nrr46l
@r_devops
Proxmox‑GitOps: Extensible GitOps (recursive Monorepo IaC, demo vid. incl.)


Proxmox‑GitOps implements a self‑contained CI/CD control plane for Proxmox VE that bootstraps from a single repository and manages itself recursively within the LXC containers under management

Repository: https://github.com/stevius10/Proxmox-GitOps
Demo (1min+): https://youtu.be/2oXDgbvFCWY?si=gSSACmVi0mO6v8xx

### Architecture
- A local bootstrap (./local/run.sh) seeds a Gitea instance and runner, initializes the pipeline, and creates an initial PR. Merging this PR transitions the system into self-management; subsequent commits converge the desired state across Proxmox LXC containers.
- The system uses a self-contained monorepo with reusable container libraries. Ansible handles provisioning against Proxmox, while Cinc (Chef) performs desired-state convergence and cross-layer orchestration where declarative modeling is insufficient.

### Concept
- Recursive self-management: the control plane executes within the managed containers to maximize reproducibility and minimize drift.
- Git as current desired state: operations map to standard Git workflows (commit, merge, rollback) in a stateless management model.
- Convention-based extensibility: add a service by copying a container definition from libs, adding a minimal cookbook and config.env; the pipeline handles provisioning, configuration, and validation.
- Loose coupling: containers remain independently replaceable and continue to function without manual follow-up.

### Environment
- Proxmox VE 8.4–9.0, Debian 13 LXC per default.
- Local bootstrap via Docker; further actions are repository-driven.

### Install
- Configure Proxmox credentials in ./local/config.json.
- Run ./local/run.sh to seed the environment.
- Accept the initial PR in the seeded Gitea instance at localhost:8080/main/config.
- Push changes to trigger provisioning, convergence, and validation on Proxmox VE.

### Trade-Offs
- The recursive bootstrap increases complexity to preserve rebuild‑from‑repo semantics and deterministic behavior.
- On Proxmox 9, stricter token privileges limit certain operations; automation uses root‑context API access where tokens are insufficient.


https://redd.it/1nrqov0
@r_devops
Deploy to production?

What's your process to go from local development to production?

I'm often using Docker on a dedicated server, but I'm curious what stuff you guys use.

Kubernetes? AWS Lambda?

https://redd.it/1nrtfbe
@r_devops
Switching from Data Science to DevOps/Cloud Engineering — need advice as a fresher

Hey everyone,

I’m a fresher who initially started preparing for Data Science, but recently I realized that almost every other person around me is going into ML/DS, and fresher entry into real Data Scientist roles is very limited (most start as Data Analysts).

After researching and discussing with mentors, I feel DevOps + Cloud Engineering suits me better since it’s more of a pure engineering role, in high demand, and has a clearer entry path for freshers. I also like the idea that later I can pivot into MLOps if I want to connect with ML.

My plan right now:

Month 1: Linux, Networking, Git, Bash/Python noscripting (+ Oracle Cloud Foundations cert in parallel)
Month 2–3: AWS/OCI core services, Docker, CI/CD, Terraform, Kubernetes basics
Month 4: Hands-on projects + cert + portfolio (GitHub)

👉 I’d love to hear from folks in the industry:

Does this switch make sense long-term compared to chasing Data Science?
For a fresher, is Cloud/DevOps a better entry point?
Any tips on what not to waste time on in the beginning?

Thanks in advance 🙏

https://redd.it/1nrqz5e
@r_devops
How do you guys handle cluster upgrades?

I am currently managing 30+ enterprise workload clusters and its upgrade time again, the clusters are mostly AWS and have 1 managed nodegrp for karpenter and other nodegroups are managed by karpenter so upgrades comparatively takes less time.

But i still have a few clusters which have self managed node groups ( some created using terraform and some using eksctl but both the terraform and the eksctl yaml is lost ) so the upgrades are hectic for these.

How do you guys handle it? Is it that you all have corresponding terraforms handy everytime or do you have some generic automation noscript written to handle such things?

If its a noscript i am also trying to write one, some advice would be much appreciated.

https://redd.it/1nrwbvy
@r_devops
New to devops, any feedback / suggestion for my IaC setup?

Hi!
I previously had a Kubernetes cluster that I was managing myself, and I decided to convert to IaC.

My setup now consists of:
\- a terraform project to bootstrap a k3s cluster on Hetzner servers, using the amazing terraform-hcloud-kube-hetzner tf module (this kinda sets up the hardware, and the really basic kubernetes resources like CNI, etc...)
\- an argocd project that manages additional resources I want available in my cluster, like cert-manager ClusterIssuer-s etc...

I think the terraform part is ok, I'm really unsure about the ArgoCD setup.
I'm new to that and it's kind of overwhelming so I have no idea whether what I'm doing is good practice.
(Also, I've read about ways to structure the repo for different environments like prod, staging, qa, etc, but since this is for my cluster which is basically a production only thing, I did not go all the way to implement that env structure)

Roast me! Here is the link to my repo: https://github.com/Giuliopime/gport

https://redd.it/1nrxvfq
@r_devops
What's your CI setup and do you like it?

Hey everyone,

I'm currently the only DevOps at my company, and I'm looking for new solutions for my CI/CD setup, as the current one is reaching its limits.
We're on GitHub action, using two self hosted runners and one remote buildkit instance. Those 3 instances are on hetzner, so disturbingly cheap. We manage around 35 users concurrency with that. We have around 300k minutes/month.
Limits of this system are obvious, concurrency is not so high, maintenance on those machines is super manual, we need to manage machines disk size etc.

What are your current setup, how many minutes do you run approximately per month, and how happy are you about your CI system?

I've looked at stuff like ARC, Phillips Terraform, blacksmith.io but they all feel like solving some issues but creating more (managing another EKS, cost high, scalability etc.)

Cheers!

https://redd.it/1nryntt
@r_devops
DevOps Colors, hows been your experience?

To provide some context to the noscript:

The idea of the DevOps can be explained to some degree with "Devs who care where "their" code runs on, and Ops that care what code runs on "their" infra" seeking the best efficiency and least issues possible. If we could assign a color based on their background there are Dev heavy Devops and Ops heavy Devops.

For example myself i'm a clear Ops person, I can hit my Scripts, IaC and Bash snippets, I know my way around K8s environment, even there is the Devside that takes me longer and probably I do it worse than some of my colleagues, thats fine we provide and split our duties.

What I've come to realize that... as time goes by more Dev-heavy folks come to the game, and I'm seen a change in how the bulk of DevOps teams approach their work, I know my code-side is weaker, and I try to think the how before thinking about code, in a sense analyze and find the easiest way or less-friction possible to ease my time in the IDE as much as possible, but with time I feel most of my colleagues jump directly to the IDE and start coding trying to find that path among VSCode lines.

Curious about others ideas, thoughts or see if my feel that Ops heavy (Think about SysAdmins, Support, Solutions Architects) in the DevOps space are becoming rarer and rarer.

https://redd.it/1ns05is
@r_devops
After Python Which Path to Choose?

I have been learning Python day and night, but now I’m confused between two areas: AI development or DevOps/Cloud.

To be honest, I don’t love either or even programming. I’m just doing it to get paid. I’m the kind of person who gets things done, even if I hate them.

So, if you were only focused on making money and solving problems at a large scale, what would you choose?

https://redd.it/1ns11pd
@r_devops
What do other people use besides kubernetes?


I began my career working directly with Kubernetes, but I’ve noticed not all companies adopt it, they often say it’s too complex. Are there real alternatives to Kubernetes? Personally, I can’t imagine managing a company’s infrastructure without it.

So what do those companies use instead to handle scaling, self-hosting, and similar needs?

https://redd.it/1ns30sw
@r_devops