How to stay updated and keep upskilling.
I have been in devops role from last 1 year. I was dealing with docker, linux machines on aws and linode. It was a small scale startup they had around >20k daily active user. I have resigned in sept as i needed a long break (4 months) due to some personal work. Currently i am a bit worried what if i forget how to do this that stuff in devops. I just wants to know how can i keep my self aligned with the market so if i start job hunting after my break i don't feel under skilled. How to practice devops on scale to keep the confidence.
Thanks
https://redd.it/1ote087
@r_devops
I have been in devops role from last 1 year. I was dealing with docker, linux machines on aws and linode. It was a small scale startup they had around >20k daily active user. I have resigned in sept as i needed a long break (4 months) due to some personal work. Currently i am a bit worried what if i forget how to do this that stuff in devops. I just wants to know how can i keep my self aligned with the market so if i start job hunting after my break i don't feel under skilled. How to practice devops on scale to keep the confidence.
Thanks
https://redd.it/1ote087
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
VOA v2.0.0 — Secrets Manager
I’ve just released VOA v2.0.0, a small open-source Secrets Manager API designed to help developers and DevOps teams securely manage and monitor sensitive data (like API keys, env vars, and credentials) across environments (dev/test/prod).
Tech stack:
FastAPI (backend)
AES encryption (secure storage)
Prometheus + Grafana (monitoring and metrics)
Dockerized setup
It’s not a big enterprise product — just a simple, educational project aimed at learning and practicing security, automation, and observability in real DevOps workflows.
🔗 GitHub repo: https://github.com/senani-derradji/VOA
you find it interesting, give it a star or share your thoughts — I’d love some feedback on what to improve or add next!
If
https://redd.it/1otf1zr
@r_devops
I’ve just released VOA v2.0.0, a small open-source Secrets Manager API designed to help developers and DevOps teams securely manage and monitor sensitive data (like API keys, env vars, and credentials) across environments (dev/test/prod).
Tech stack:
FastAPI (backend)
AES encryption (secure storage)
Prometheus + Grafana (monitoring and metrics)
Dockerized setup
It’s not a big enterprise product — just a simple, educational project aimed at learning and practicing security, automation, and observability in real DevOps workflows.
🔗 GitHub repo: https://github.com/senani-derradji/VOA
you find it interesting, give it a star or share your thoughts — I’d love some feedback on what to improve or add next!
If
https://redd.it/1otf1zr
@r_devops
GitHub
GitHub - senani-derradji/VOA: VOA (VaulityOpsAPI) is a FastAPI-based secrets management platform for DevOps. Securely store, retrieve…
VOA (VaulityOpsAPI) is a FastAPI-based secrets management platform for DevOps. Securely store, retrieve, and audit environment variables, API keys, and passwords across dev, staging, and prod envir...
In 2022, I wrote that DevOps had become waste, in 2025 AI is the new waste!
In 2022, I said DevOps had become waste.
The response?
"DevOps can't be waste we need automation!"
They missed the point.
DevOps principles were right.
But when every team rebuilds the same CI/CD pipelines, writes the same Terraform modules, and solves the same problems in isolation
that’s not DevOps.
That’s local optimization at scale.
Now it’s 2025. AI is the new waste.
Team A spends two sprints wiring up Claude to “understand” their codebase.
They chunk it, inject docs, tweak prompts.
Team B? Doing the same thing.
Different team. Same half-baked playbook.
No shared learning. No standardization. No outcomes tracked.
And most orgs?
Still stuck trying to pick Copilot vs. CodeWhisperer vs. Windsurf
with zero plan to measure impact or build repeatable systems.
This is Jenkins sprawl all over again but for cognition.
I call the fix: OutcomeOps
https://www.outcomeops.ai/blogs/outcomeops-ai-is-the-new-waste
https://redd.it/1oti8e8
@r_devops
In 2022, I said DevOps had become waste.
The response?
"DevOps can't be waste we need automation!"
They missed the point.
DevOps principles were right.
But when every team rebuilds the same CI/CD pipelines, writes the same Terraform modules, and solves the same problems in isolation
that’s not DevOps.
That’s local optimization at scale.
Now it’s 2025. AI is the new waste.
Team A spends two sprints wiring up Claude to “understand” their codebase.
They chunk it, inject docs, tweak prompts.
Team B? Doing the same thing.
Different team. Same half-baked playbook.
No shared learning. No standardization. No outcomes tracked.
And most orgs?
Still stuck trying to pick Copilot vs. CodeWhisperer vs. Windsurf
with zero plan to measure impact or build repeatable systems.
This is Jenkins sprawl all over again but for cognition.
I call the fix: OutcomeOps
https://www.outcomeops.ai/blogs/outcomeops-ai-is-the-new-waste
https://redd.it/1oti8e8
@r_devops
www.outcomeops.ai
OutcomeOps: AI Is the New Waste - OutcomeOps Blog
In 2022, I wrote that DevOps had become waste. Now, in 2025, AI local optimization is the new waste—thousands of teams rebuilding the same RAG systems, prompts, and context pipelines in isolation.
CKA Preparation
Im preparing for the CKA Cert. I already did these courses: LFS158 & LFS258, and I’m administering the k8s cluster of my company for a little more then a year now on pretty much a daily basis. I did the killerkoda tests & also did both of the killer.sh mock exams. In the first mock exam, I only scored about 50% and in the second one even worse. I used the 120min timer to make the test as realistic as possible. After this I redid all of the answers that I failed on & got 100% correct. I didn’t really have issues with specific topics, my only problem was the time constraint.
So my question: Am I prepared enough, even though I technically failed the mock exams? I read that killer.sh exams are much harder then the real exam. If that’s not true, I don’t really know how to better prepare for the exam, because I prepared using all of the resources that I’m aware of.
Thanks :)
https://redd.it/1otedcw
@r_devops
Im preparing for the CKA Cert. I already did these courses: LFS158 & LFS258, and I’m administering the k8s cluster of my company for a little more then a year now on pretty much a daily basis. I did the killerkoda tests & also did both of the killer.sh mock exams. In the first mock exam, I only scored about 50% and in the second one even worse. I used the 120min timer to make the test as realistic as possible. After this I redid all of the answers that I failed on & got 100% correct. I didn’t really have issues with specific topics, my only problem was the time constraint.
So my question: Am I prepared enough, even though I technically failed the mock exams? I read that killer.sh exams are much harder then the real exam. If that’s not true, I don’t really know how to better prepare for the exam, because I prepared using all of the resources that I’m aware of.
Thanks :)
https://redd.it/1otedcw
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
How do you check or enforce code documentation in your pipelines (C/C++ & Python)?
Hey,
Currently working on improving how we enforce code documentation coverage across a few repositories, and I’d love to hear how others handle this.
We have three main repos:
one in C++
one in C and C++
one in Python
For C and C++, we’re using Doxygen with Javadoc-style comments.
For Python, we use Google-style docstrings.
Right now, for the C and C++ part, we have a CI pipeline that runs Doxygen for every merge request and compares the documentation coverage against the main branch. If coverage decreases, the user gets notified, and the MR is blocked.
That works okay, but I’m wondering:
Are there better or existing tools or CI integrations that already handle documentation checks like this? Only Open source and applying locally would be fine.
What would be a good equivalent setup for Python? (e.g., something to validate or measure docstring coverage)
Has anyone implemented pre-commit or pre-push git hooks that check for missing documentation or docstring issues before the MR even gets created?
Thanks in advance!
https://redd.it/1otjf9r
@r_devops
Hey,
Currently working on improving how we enforce code documentation coverage across a few repositories, and I’d love to hear how others handle this.
We have three main repos:
one in C++
one in C and C++
one in Python
For C and C++, we’re using Doxygen with Javadoc-style comments.
For Python, we use Google-style docstrings.
Right now, for the C and C++ part, we have a CI pipeline that runs Doxygen for every merge request and compares the documentation coverage against the main branch. If coverage decreases, the user gets notified, and the MR is blocked.
That works okay, but I’m wondering:
Are there better or existing tools or CI integrations that already handle documentation checks like this? Only Open source and applying locally would be fine.
What would be a good equivalent setup for Python? (e.g., something to validate or measure docstring coverage)
Has anyone implemented pre-commit or pre-push git hooks that check for missing documentation or docstring issues before the MR even gets created?
Thanks in advance!
https://redd.it/1otjf9r
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Open-Source ACME server - 100% CertBot compatible - One binary
Hi everyone!
We have developed an Acme server for our use case. It is written in Rust, which means you only need to work with a single binary. In file mode, our test is 100% compatible with the existing Certbot solution.
For more details, visit: https://github.com/arxignis/ssl-storage
**Summary:**
✅ Written in Rust
✅ Fully compatible with Certbot
✅ Utilizes a Redis backend for storage
✅ Supports distributed mode (with Redis)
✅ 100% compatible with CertBot
✅ Redis backend as storage mode
✅ Distributed mode (with Redis)
https://redd.it/1othpt8
@r_devops
Hi everyone!
We have developed an Acme server for our use case. It is written in Rust, which means you only need to work with a single binary. In file mode, our test is 100% compatible with the existing Certbot solution.
For more details, visit: https://github.com/arxignis/ssl-storage
**Summary:**
✅ Written in Rust
✅ Fully compatible with Certbot
✅ Utilizes a Redis backend for storage
✅ Supports distributed mode (with Redis)
✅ 100% compatible with CertBot
✅ Redis backend as storage mode
✅ Distributed mode (with Redis)
https://redd.it/1othpt8
@r_devops
GitHub
GitHub - arxignis/ssl-storage: Distributed ACME SSL storage
Distributed ACME SSL storage. Contribute to arxignis/ssl-storage development by creating an account on GitHub.
Browsing helm chart from terminal - LazyHelm
Hi community!
Sometimes, when I deploy or test some application, I prefer looking into helm charts using directly the terminal and I found using helm commands alone can get a bit tedious, so I tried to created something to make it easier.
So I tried to create (with ai helps) something that makes the process easier, LazyHelm.
It’s a small personal project I built to make my own workflow smoother, but I hope it might help someone else too.
What it does:
Organized menu system to browse local repositories or search Artifact Hub
Browse your configured Helm repos and discover all available charts
Find charts across Artifact Hub directly from the terminal
Add, remove, and update repository indexes with simple keystrokes
Inspect chart values with syntax highlighting and diff between versions
Modify values in your preferred editor ($EDITOR) with YAML validation
Fuzzy search through repositories, charts, and values
Copy YAML paths to clipboard or export values to files
All in your terminal. No need to remember helm commands or manually fetch values.
Installation via Homebrew:
You can install LazyHelm using Homebrew:
- brew install alessandropitocchi/lazyhelm/lazyhelm
GitHub: https://github.com/alessandropitocchi/lazyhelm
Any feedback, suggestions, or feature requests are very welcome!
Thanks for reading!
https://redd.it/1otopeq
@r_devops
Hi community!
Sometimes, when I deploy or test some application, I prefer looking into helm charts using directly the terminal and I found using helm commands alone can get a bit tedious, so I tried to created something to make it easier.
So I tried to create (with ai helps) something that makes the process easier, LazyHelm.
It’s a small personal project I built to make my own workflow smoother, but I hope it might help someone else too.
What it does:
Organized menu system to browse local repositories or search Artifact Hub
Browse your configured Helm repos and discover all available charts
Find charts across Artifact Hub directly from the terminal
Add, remove, and update repository indexes with simple keystrokes
Inspect chart values with syntax highlighting and diff between versions
Modify values in your preferred editor ($EDITOR) with YAML validation
Fuzzy search through repositories, charts, and values
Copy YAML paths to clipboard or export values to files
All in your terminal. No need to remember helm commands or manually fetch values.
Installation via Homebrew:
You can install LazyHelm using Homebrew:
- brew install alessandropitocchi/lazyhelm/lazyhelm
GitHub: https://github.com/alessandropitocchi/lazyhelm
Any feedback, suggestions, or feature requests are very welcome!
Thanks for reading!
https://redd.it/1otopeq
@r_devops
GitHub
GitHub - alessandropitocchi/lazyhelm
Contribute to alessandropitocchi/lazyhelm development by creating an account on GitHub.
QA team was cut in half, facing the same release pressure. thoughts?
we lost half of our QA team in the last round of budget cuts, but somehow leadership is still expecting us to keep shipping every 2 weeks. I mean manual regression alone takes most of the sprint, not to mention the pain of cross device tests as we're testing across web + android.
the team is already burned out and lacks resources now, higher ups say we can fix this with automation but setting up new frameworks feels like starting a new project and we can't afford to waste any more time experimenting nor do we have the engineering bandwidth now...
has anyone successfully automated testing across devices without hiring more engineers? AI tools? Low-code? we need something good and we need it SOON...
https://redd.it/1otts8b
@r_devops
we lost half of our QA team in the last round of budget cuts, but somehow leadership is still expecting us to keep shipping every 2 weeks. I mean manual regression alone takes most of the sprint, not to mention the pain of cross device tests as we're testing across web + android.
the team is already burned out and lacks resources now, higher ups say we can fix this with automation but setting up new frameworks feels like starting a new project and we can't afford to waste any more time experimenting nor do we have the engineering bandwidth now...
has anyone successfully automated testing across devices without hiring more engineers? AI tools? Low-code? we need something good and we need it SOON...
https://redd.it/1otts8b
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Moving to a mid level position
Hey all,
So, I've been within the devops/platform engineering space for just under 2 years now. I come from a non tech background but I'm firmly in the tech space now.
But I wanted to understand how can I make that move from junior to mid level engineer? I have a good solid grasp of Terraform, GitLab CI. Some Docker and K8s skills (fairly new for a project on EKS). My main cloud is AWS for the past 3 years. I'm currently also getting involved with some other clouds like oci.
But I feel like I don't have a strong understanding of some basic stuff that an IT or tech guy should have. Networking skills are probably lacking tbh. I'd love to increase my security skills also.
I would love to have someone as a mentor to help guide and advise me through this process.
https://redd.it/1otu4z2
@r_devops
Hey all,
So, I've been within the devops/platform engineering space for just under 2 years now. I come from a non tech background but I'm firmly in the tech space now.
But I wanted to understand how can I make that move from junior to mid level engineer? I have a good solid grasp of Terraform, GitLab CI. Some Docker and K8s skills (fairly new for a project on EKS). My main cloud is AWS for the past 3 years. I'm currently also getting involved with some other clouds like oci.
But I feel like I don't have a strong understanding of some basic stuff that an IT or tech guy should have. Networking skills are probably lacking tbh. I'd love to increase my security skills also.
I would love to have someone as a mentor to help guide and advise me through this process.
https://redd.it/1otu4z2
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Policy as Code
I recently moved our company’s azure policy away from being manual process through the azure web portal to a pipeline using terraform. It’s working but it’s not great, I’m wondering how others manage their Azure Policy, or AWS scps
https://redd.it/1otyhkh
@r_devops
I recently moved our company’s azure policy away from being manual process through the azure web portal to a pipeline using terraform. It’s working but it’s not great, I’m wondering how others manage their Azure Policy, or AWS scps
https://redd.it/1otyhkh
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
HTTP Parameter Pollution: Making Servers Disagree on What You Sent 🔀
https://instatunnel.my/blog/http-parameter-pollution-making-servers-disagree-on-what-you-sent
https://redd.it/1otzhyy
@r_devops
https://instatunnel.my/blog/http-parameter-pollution-making-servers-disagree-on-what-you-sent
https://redd.it/1otzhyy
@r_devops
InstaTunnel
HTTP Parameter Pollution (HPP): When Servers Disagree on You
Discover how HTTP Parameter Pollution (HPP) uses duplicate or conflicting parameters to confuse servers, bypass WAFs and filters, and enable data leakage
Tools Auto tagging
So I found a cool project called Yor by paloalto that does some great tagging automation.
Sadly project looks dead, docs are lacking, and it doesn't support OpenTofu.
Are there any other tools like this out there, that are actively maintained?
Looking for automating, git repo and project tags at a minimum.
https://redd.it/1otxnpf
@r_devops
So I found a cool project called Yor by paloalto that does some great tagging automation.
Sadly project looks dead, docs are lacking, and it doesn't support OpenTofu.
Are there any other tools like this out there, that are actively maintained?
Looking for automating, git repo and project tags at a minimum.
https://redd.it/1otxnpf
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Tools for solo PMs or very small PM teams?
Working as the only PM at a small startup and most PM tools feel like overkill. What do other solo PMs use that's not overly complicated but still helps stay organized?
https://redd.it/1ou0hwo
@r_devops
Working as the only PM at a small startup and most PM tools feel like overkill. What do other solo PMs use that's not overly complicated but still helps stay organized?
https://redd.it/1ou0hwo
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
We at SigNoz shipped the 100th release of our open-source observability platform
When we started SigNoz, we wanted to build an "open" observability platform:
Open source
Based on OpenTelemetry
Self-host it in your infra if needed
All in one, with transparent pricing that doesn't punish you for actually using your monitoring tool.
v0.100.0 adds:
Span percentiles \- catch performance outliers in your traces without drowning in data
Infrastructure metrics in traces \- correlate app performance with resource usage
Cost meter alerts \- track your observability spend so you're not hit with surprise bills
Full changelog: https://signoz.io/changelog/
We're not trying to replace everything overnight, but if you're tired of vendor lock-in or paying per-host nonsense, might be worth a look :)
GitHub: https://github.com/SigNoz/signoz
https://redd.it/1ou4t81
@r_devops
When we started SigNoz, we wanted to build an "open" observability platform:
Open source
Based on OpenTelemetry
Self-host it in your infra if needed
All in one, with transparent pricing that doesn't punish you for actually using your monitoring tool.
v0.100.0 adds:
Span percentiles \- catch performance outliers in your traces without drowning in data
Infrastructure metrics in traces \- correlate app performance with resource usage
Cost meter alerts \- track your observability spend so you're not hit with surprise bills
Full changelog: https://signoz.io/changelog/
We're not trying to replace everything overnight, but if you're tired of vendor lock-in or paying per-host nonsense, might be worth a look :)
GitHub: https://github.com/SigNoz/signoz
https://redd.it/1ou4t81
@r_devops
SigNoz
SigNoz is an open-source observability tool powered by OpenTelemetry. Get APM, logs, traces, metrics, exceptions, & alerts in a single tool.
Kodekloud Black Friday sales
I recall seeing the similar pricing and discount as regular days, am I missing something to apply the discount code for annual sub on this sales?
https://redd.it/1ou5xwl
@r_devops
I recall seeing the similar pricing and discount as regular days, am I missing something to apply the discount code for annual sub on this sales?
https://redd.it/1ou5xwl
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Do your tools ever slowly stop reflecting what's actually happening?
Something I keep running into is that we set up the perfect board, workflows, dashboards, all of it and then two weeks later it’s already out of sync with reality. The plan and the actual work just start drifting apart. Tickets stay “in progress” when they’re blocked. Priorities shift but the board doesn’t. People share updates in side conversations that never make it back into the system.
It’s not that the tools are bad. We’ve tried Jira, ClickUp, even some of the more visual platforms. They all work at first. The real problem seems to be keeping things up-to-date once things get messy and priorities move. And that’s exactly when the visibility would matter the most.
So I’m wondering, how do you keep your source of truth accurate when the work is constantly changing? Is it the tool? The rituals? The culture?
https://redd.it/1ou6ae5
@r_devops
Something I keep running into is that we set up the perfect board, workflows, dashboards, all of it and then two weeks later it’s already out of sync with reality. The plan and the actual work just start drifting apart. Tickets stay “in progress” when they’re blocked. Priorities shift but the board doesn’t. People share updates in side conversations that never make it back into the system.
It’s not that the tools are bad. We’ve tried Jira, ClickUp, even some of the more visual platforms. They all work at first. The real problem seems to be keeping things up-to-date once things get messy and priorities move. And that’s exactly when the visibility would matter the most.
So I’m wondering, how do you keep your source of truth accurate when the work is constantly changing? Is it the tool? The rituals? The culture?
https://redd.it/1ou6ae5
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Offered 6LPA at a 5-year-old startup (3-month notice) — Accept or wait?
hey guys,
I got a full-time DevOps offer after my internship, 6 LPA package(Remote). The only catch is a 3-month notice period .Not getting many interview calls lately, but I’m worried this might limit my growth or make switching tougher later. Do you think it’s better to take it for now and gain some experience, or hold out for something around 7–8 LPA?
Would love to hear what others did in a similar situation.
https://redd.it/1ouaqy1
@r_devops
hey guys,
I got a full-time DevOps offer after my internship, 6 LPA package(Remote). The only catch is a 3-month notice period .Not getting many interview calls lately, but I’m worried this might limit my growth or make switching tougher later. Do you think it’s better to take it for now and gain some experience, or hold out for something around 7–8 LPA?
Would love to hear what others did in a similar situation.
https://redd.it/1ouaqy1
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
used ai for monolith to microservices migration. saved maybe 20% on configs, zero help on the actual hard parts
just wrapped up migrating our 80k line monolith to microservices. 5 months with 3 devops + 4 backend devs.
figured id try ai tools since everyones hyping them. mixed bag honestly.
stuff that actually helped:
k8s configs - copilot spit out decent yaml. still had to fix half of it but beat writing from scratch.
ci/cd pipelines - chatgpt gave me basic github actions structure. we added our deploy logic on top.
dockerfiles - claude suggested multi stage builds i hadnt used before. learned something new.
task planning - tried verdent and cursor for breaking down the migration phases. cursor gave me a list of steps but verdent actually showed dependencies between tasks and what order made sense. like it caught that we needed to set up the message queue before splitting the order service. helped us not miss steps for the complex services.
terraform modules - copilot again. generated basic module structure.
stuff that was useless:
service boundaries - ai suggested some boundaries based on data models. we obviously knew better but still spent 3 weeks with the team figuring out actual domain boundaries based on business logic.
data migration - kept suggesting saga pattern but didnt understand our constraints with payment processing. ended up doing event sourcing with phased rollout. ai had zero clue about our actual requirements.
observability - generated basic prometheus stuff but didnt understand our actual metrics or what we should alert on.
numbers:
estimated 6 months, took 5
ai probably saved 2-3 weeks on config and planning work
infrastructure costs up 40% tho (ai never mentioned that)
worst part was ai saying to migrate payment service all at once with feature flags. we do high volume transactions, cant risk that. took 3 weeks doing strangler pattern instead.
now we got 12 services, 10 in prod. still migrating the last 2 (reporting and analytics). deploying went from 45min for the whole monolith to 8min for whatever service changed. nice since we usually only touch 1-2 services anyway.
but distributed tracing is a pain now. more stuff to monitor, network latency issues, eventual consistency headaches. ai was zero help with any of that.
so yeah. ai good for boring config stuff. completely useless for actual architecture decisions. distributed systems are still hard.
anyone else migrate recently? what worked for you
https://redd.it/1oub853
@r_devops
just wrapped up migrating our 80k line monolith to microservices. 5 months with 3 devops + 4 backend devs.
figured id try ai tools since everyones hyping them. mixed bag honestly.
stuff that actually helped:
k8s configs - copilot spit out decent yaml. still had to fix half of it but beat writing from scratch.
ci/cd pipelines - chatgpt gave me basic github actions structure. we added our deploy logic on top.
dockerfiles - claude suggested multi stage builds i hadnt used before. learned something new.
task planning - tried verdent and cursor for breaking down the migration phases. cursor gave me a list of steps but verdent actually showed dependencies between tasks and what order made sense. like it caught that we needed to set up the message queue before splitting the order service. helped us not miss steps for the complex services.
terraform modules - copilot again. generated basic module structure.
stuff that was useless:
service boundaries - ai suggested some boundaries based on data models. we obviously knew better but still spent 3 weeks with the team figuring out actual domain boundaries based on business logic.
data migration - kept suggesting saga pattern but didnt understand our constraints with payment processing. ended up doing event sourcing with phased rollout. ai had zero clue about our actual requirements.
observability - generated basic prometheus stuff but didnt understand our actual metrics or what we should alert on.
numbers:
estimated 6 months, took 5
ai probably saved 2-3 weeks on config and planning work
infrastructure costs up 40% tho (ai never mentioned that)
worst part was ai saying to migrate payment service all at once with feature flags. we do high volume transactions, cant risk that. took 3 weeks doing strangler pattern instead.
now we got 12 services, 10 in prod. still migrating the last 2 (reporting and analytics). deploying went from 45min for the whole monolith to 8min for whatever service changed. nice since we usually only touch 1-2 services anyway.
but distributed tracing is a pain now. more stuff to monitor, network latency issues, eventual consistency headaches. ai was zero help with any of that.
so yeah. ai good for boring config stuff. completely useless for actual architecture decisions. distributed systems are still hard.
anyone else migrate recently? what worked for you
https://redd.it/1oub853
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Hi, is there here anyone configured gitlab cicd pipelines for OCI terraform ?
I am facing issues and need help from someone who did it already for OCI (Oracle Cloud)
https://redd.it/1ouamkg
@r_devops
I am facing issues and need help from someone who did it already for OCI (Oracle Cloud)
https://redd.it/1ouamkg
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Coroot 1.17 - FOSS, self-hosted, eBPF-powered observability now has multi-cluster support
For new users: Coroot is an Apache 2.0 open source observability tool designed to help developers quickly find and resolve the root cause of incidents. With eBPF, the Coroot node agent automatically visualizes logs, metrics, profiles, spans, traces, a map of your services, and suggests tips on reducing cloud costs. Compatible with Prometheus, Clickhouse, VictoriaMetrics, OTEL, and all your other favourite FOSS usual suspects.
We’ve had a couple major updates recently to include multi-cluster and OTEL/gRPC support. A multi-cluster Coroot project can help simplify and unify monitoring for applications deployed across multiple kubernetes clusters, regions, or data centers (without duplicating ingestion pipelines.) Additionally, OTEL/gRPC compatibility can help make the tool more efficient for users who depend on high-volume data transfers.
Feedback is always welcome to help improve open observability for everyone, so give us a nudge with any bug reports or questions.
https://redd.it/1ouf3l5
@r_devops
For new users: Coroot is an Apache 2.0 open source observability tool designed to help developers quickly find and resolve the root cause of incidents. With eBPF, the Coroot node agent automatically visualizes logs, metrics, profiles, spans, traces, a map of your services, and suggests tips on reducing cloud costs. Compatible with Prometheus, Clickhouse, VictoriaMetrics, OTEL, and all your other favourite FOSS usual suspects.
We’ve had a couple major updates recently to include multi-cluster and OTEL/gRPC support. A multi-cluster Coroot project can help simplify and unify monitoring for applications deployed across multiple kubernetes clusters, regions, or data centers (without duplicating ingestion pipelines.) Additionally, OTEL/gRPC compatibility can help make the tool more efficient for users who depend on high-volume data transfers.
Feedback is always welcome to help improve open observability for everyone, so give us a nudge with any bug reports or questions.
https://redd.it/1ouf3l5
@r_devops
GitHub
GitHub - coroot/coroot: Coroot is an open-source observability and APM tool with AI-powered Root Cause Analysis. It combines metrics…
Coroot is an open-source observability and APM tool with AI-powered Root Cause Analysis. It combines metrics, logs, traces, continuous profiling, and SLO-based alerting with predefined dashboards a...
65% of Startups from Forbes AI 50 Leaked Secrets on GitHub
https://www.wiz.io/blog/forbes-ai-50-leaking-secrets
https://redd.it/1ouii3y
@r_devops
https://www.wiz.io/blog/forbes-ai-50-leaking-secrets
https://redd.it/1ouii3y
@r_devops
wiz.io
65% of Startups from Forbes AI 50 Leaked Secrets on GitHub | Wiz Blog
A Wiz investigation into the Forbes AI 50 reveals 65% of leading AI startups had leaked secrets. See real examples, leak types, and how to prevent this.