AI is a Corporate Fad where I work
The noscript says it all. In my workplace (big company) we have non-technical decision makers asking for integrations of technology that they don't understand with existing technologies that they don't understand. What could go wrong financially?
My only hope is that this fad replaces the existing fad of hiring swaths of inexpensive out of town engineers to provide "top notch" solution design that falls flat at the implementation phase.
What's your experience?
https://redd.it/1onilgi
@r_devops
The noscript says it all. In my workplace (big company) we have non-technical decision makers asking for integrations of technology that they don't understand with existing technologies that they don't understand. What could go wrong financially?
My only hope is that this fad replaces the existing fad of hiring swaths of inexpensive out of town engineers to provide "top notch" solution design that falls flat at the implementation phase.
What's your experience?
https://redd.it/1onilgi
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Gprxy: Go based SSO-first, psql-compatible proxy
https://github.com/sathwick-p/gprxy
Hey all,
I built a postgresql proxy for AWS RDS, the reason i wrote this is because the current way to access and run queries on RDS is via having db users and in bigger organization it is impractical to have multiple db users for each user/team, and yes even IAM authentication exists for this same reason in RDS i personally did not find it the best way to use as it would required a bunch of configuration and changes in the RDS.
The idea here is by connecting via this proxy you would just have to run the login command that would let you do a SSO based login which will authenticate you through an IDP like azure AD before connecting to the db. Also helps me with user level audit logs
I had been looking for an opensource solution but could not find any hence rolled out my own, currently deployed and being used via k8s
Please check it out and let me know if you find it useful or have feedback, I’d really appreciate hearing from y'all.
Thanks!
https://redd.it/1oni3df
@r_devops
https://github.com/sathwick-p/gprxy
Hey all,
I built a postgresql proxy for AWS RDS, the reason i wrote this is because the current way to access and run queries on RDS is via having db users and in bigger organization it is impractical to have multiple db users for each user/team, and yes even IAM authentication exists for this same reason in RDS i personally did not find it the best way to use as it would required a bunch of configuration and changes in the RDS.
The idea here is by connecting via this proxy you would just have to run the login command that would let you do a SSO based login which will authenticate you through an IDP like azure AD before connecting to the db. Also helps me with user level audit logs
I had been looking for an opensource solution but could not find any hence rolled out my own, currently deployed and being used via k8s
Please check it out and let me know if you find it useful or have feedback, I’d really appreciate hearing from y'all.
Thanks!
https://redd.it/1oni3df
@r_devops
GitHub
GitHub - sathwick-p/gprxy: Go based SSO-first, psql-compatible proxy
Go based SSO-first, psql-compatible proxy. Contribute to sathwick-p/gprxy development by creating an account on GitHub.
Just got $5K AWS credits approved for my startup
Didn’t expect this to still work in 2025, but I just got **$5,000 in AWS credits** approved for my small startup.
We’re not in YC or any accelerator just a verified startup with:
* a **website**
* a **business email**
* and an actual product in progress
It took around 2–3 days to get verified, and the credits were added directly to the AWS account.
So if you’re building something and have your own domain, there’s still a valid path to get AWS credits even if you’re not part of Activate.
If anyone’s curious or wants to check if they’re eligible, DM me I can share the steps.
https://redd.it/1onmg20
@r_devops
Didn’t expect this to still work in 2025, but I just got **$5,000 in AWS credits** approved for my small startup.
We’re not in YC or any accelerator just a verified startup with:
* a **website**
* a **business email**
* and an actual product in progress
It took around 2–3 days to get verified, and the credits were added directly to the AWS account.
So if you’re building something and have your own domain, there’s still a valid path to get AWS credits even if you’re not part of Activate.
If anyone’s curious or wants to check if they’re eligible, DM me I can share the steps.
https://redd.it/1onmg20
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Migrating from Octopus Deploy to Gitlab. What are Pros and Cons?
Due to reasons I won't get into, we might need to move from Octopus Deploy to Gitlab for CICD. Trying to come up with some pros and cons so I can convince management to keep Octopus (despite the cost). Here are some of pros for having Octopus that I have listed:
Release management.
If we need to roll back to a previously functioning version of our code, we can simply click on the previous release and then leisurely work on fixing the problem. (sometimes issues aren't always visible in QA or Staging). Gitlab doesn't seem to have this.
Script Console
Octopus lets us send commands (eg, iisreset) to an entire batch of VMs in one shot instead having to write something that would loop through a list of VMs, or God forbid, remoting into each VM manually. GitLab doesn't seem to have that either. This comes in really handy when we need to quickly run a task in the middle of an outage.
Variable Management and Substitution
Scoping variable with different values seem to be handled much better in Octopus compared to GitLab. Also I could not find anything that says you can do variable substitution in your code for files like .config, .json files. No .NET variable substitution either in Gitlab.
Pipeline Design
Gitlab pipeline seems to be all YAML which means a lot of the tasks that Octo does for you, like IIS configuration, Kubernetes deployments, etc., will have to noscripted from scratch. (Correct me if I'm wrong on this).
These some of the Pros of Octopus I could think of. Are there any more I can use to back up my argument.
Also is there anyone who went through the same exercise? What is your experience using Gitlab after having Octopus for a while?
https://redd.it/1onlv3s
@r_devops
Due to reasons I won't get into, we might need to move from Octopus Deploy to Gitlab for CICD. Trying to come up with some pros and cons so I can convince management to keep Octopus (despite the cost). Here are some of pros for having Octopus that I have listed:
Release management.
If we need to roll back to a previously functioning version of our code, we can simply click on the previous release and then leisurely work on fixing the problem. (sometimes issues aren't always visible in QA or Staging). Gitlab doesn't seem to have this.
Script Console
Octopus lets us send commands (eg, iisreset) to an entire batch of VMs in one shot instead having to write something that would loop through a list of VMs, or God forbid, remoting into each VM manually. GitLab doesn't seem to have that either. This comes in really handy when we need to quickly run a task in the middle of an outage.
Variable Management and Substitution
Scoping variable with different values seem to be handled much better in Octopus compared to GitLab. Also I could not find anything that says you can do variable substitution in your code for files like .config, .json files. No .NET variable substitution either in Gitlab.
Pipeline Design
Gitlab pipeline seems to be all YAML which means a lot of the tasks that Octo does for you, like IIS configuration, Kubernetes deployments, etc., will have to noscripted from scratch. (Correct me if I'm wrong on this).
These some of the Pros of Octopus I could think of. Are there any more I can use to back up my argument.
Also is there anyone who went through the same exercise? What is your experience using Gitlab after having Octopus for a while?
https://redd.it/1onlv3s
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Curious how folks are feeling regarding ethics in the current political climate with regards to tech?
I'm asking the question in this sub on the basis that people have to have a reasonable level of experience to be in this field across disciplines. (I did the helpdesk - Sysad - DevOps route, for context I started that journey in the late 90s).
If not allowed (I understand politics is a sensitive subject) I will completely understand mods removing the post.
I'm between contracts at the moment, and in the last decade at least, whenever I'm not working I get offers to work for "gaming" (gambling) sites that I always turn down... I have a number of friends who wrecked their lives through gambling addiction - I wouldn't feel comfortable taking that paycheck.
(I'm not shitting on people who do, I get it. No judgement... Just a personal thing on my part).
I recently watched a pretty in-depth breakdown of the facial recognition AI stuff being trialled in the US (additional context, I'm not American nor do I live there anymore, but it's fair to assume it's coming everywhere soon), but I have previously worked for a number of US companies, including at least one that I know is involved in stuff that makes me feel pretty uncomfortable about the way the technology is progressing, and importantly, the things it is being used for.
I suppose I'm speaking more to the greybeards and grey hats in this community - but I'm curious to gauge how folks are feeling about developing and supporting this kind of thing?
https://redd.it/1onpv8x
@r_devops
I'm asking the question in this sub on the basis that people have to have a reasonable level of experience to be in this field across disciplines. (I did the helpdesk - Sysad - DevOps route, for context I started that journey in the late 90s).
If not allowed (I understand politics is a sensitive subject) I will completely understand mods removing the post.
I'm between contracts at the moment, and in the last decade at least, whenever I'm not working I get offers to work for "gaming" (gambling) sites that I always turn down... I have a number of friends who wrecked their lives through gambling addiction - I wouldn't feel comfortable taking that paycheck.
(I'm not shitting on people who do, I get it. No judgement... Just a personal thing on my part).
I recently watched a pretty in-depth breakdown of the facial recognition AI stuff being trialled in the US (additional context, I'm not American nor do I live there anymore, but it's fair to assume it's coming everywhere soon), but I have previously worked for a number of US companies, including at least one that I know is involved in stuff that makes me feel pretty uncomfortable about the way the technology is progressing, and importantly, the things it is being used for.
I suppose I'm speaking more to the greybeards and grey hats in this community - but I'm curious to gauge how folks are feeling about developing and supporting this kind of thing?
https://redd.it/1onpv8x
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
The APM paradox
I've recently been thinking about how to get more developers (especially on smaller teams) to adopt observability practices, and put together some thoughts about how we're approaching it at the monitoring tool I'm building. We're a small team of developers who have been on-call for critical infrastructure for the past 13 years, and have found that while "APM" tools tend to be more developer-focused, we've generally found logging to be more essential for our own systems (which led us to build a structured logging tool that encourages wide events).
I'm curious what y'all think — how can we encourage more developers to learn about observability?
https://www.honeybadger.io/blog/apm-paradox/
https://redd.it/1onmrnj
@r_devops
I've recently been thinking about how to get more developers (especially on smaller teams) to adopt observability practices, and put together some thoughts about how we're approaching it at the monitoring tool I'm building. We're a small team of developers who have been on-call for critical infrastructure for the past 13 years, and have found that while "APM" tools tend to be more developer-focused, we've generally found logging to be more essential for our own systems (which led us to build a structured logging tool that encourages wide events).
I'm curious what y'all think — how can we encourage more developers to learn about observability?
https://www.honeybadger.io/blog/apm-paradox/
https://redd.it/1onmrnj
@r_devops
Honeybadger Developer Blog
The APM paradox: Too much data, too few answers
Most dev teams don't need 47 dashboards or petabytes of logs—they need answers at 2 AM. Explore the evolution from APM to observability.
Why do cron monitors act like a job "running" = "working"?
Most cron monitors are useless if the job executes but doesn't do what it's supposed to. I don't care if the noscript ran. I care if:
- it returned an error
- it output nothing
- it took 10x longer than usual
- it "succeeded" but wrote an empty file
All I get is "✓ ping received" like everything's fine.
Anything out there that actually checks exit status, runtime anomalies, or output sanity? Or does everyone just build this crap themselves?
https://redd.it/1onrwrl
@r_devops
Most cron monitors are useless if the job executes but doesn't do what it's supposed to. I don't care if the noscript ran. I care if:
- it returned an error
- it output nothing
- it took 10x longer than usual
- it "succeeded" but wrote an empty file
All I get is "✓ ping received" like everything's fine.
Anything out there that actually checks exit status, runtime anomalies, or output sanity? Or does everyone just build this crap themselves?
https://redd.it/1onrwrl
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Custom Podman Container Dashboard?
I have a bunch of docker containers(well technically podman containers) running on a Linux node and its getting to a point where its annoying to keep a track of all the containers. I have all the necessary identifying information(like requestor, poc etc.) added as labels to each container.
I'm looking for a way to create something like a dashboard to present this information like Container name, status, label1, label2, label3 in a nice tabular form.
I've already experimented with Portainer and Cockpit but couldn't really create a customized view per my needs. Does anyone have any ideas?
https://redd.it/1onsszc
@r_devops
I have a bunch of docker containers(well technically podman containers) running on a Linux node and its getting to a point where its annoying to keep a track of all the containers. I have all the necessary identifying information(like requestor, poc etc.) added as labels to each container.
I'm looking for a way to create something like a dashboard to present this information like Container name, status, label1, label2, label3 in a nice tabular form.
I've already experimented with Portainer and Cockpit but couldn't really create a customized view per my needs. Does anyone have any ideas?
https://redd.it/1onsszc
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
How do you size VPS resources for different kinds of websites? Looking for real-world experience and examples.
I’m trying to understand how to estimate VPS resource requirements for different kinds of websites — not just from theory, but based on real-world experience.
Are there any guidelines or rules of thumb you use (or a guide you’d recommend) for deciding how much CPU, RAM, and disk to allocate depending on things like:
* Average daily concurrent visitors
* Site complexity (static site → lightweight web app → high-load dynamic site)
* Whether a database is used and how large it is
* Whether caching or CDN layers are implemented
I know “it depends” — but I’d really like to hear from people who’ve done capacity planning for real sites:
What patterns or lessons did you learn?
* What setups worked well or didn’t?
* Any sample configurations you can share (e.g., “For a small Django app with \~10k daily visitors and caching, we used 2 vCPUs and 4 GB RAM with good performance.”)?
I’m mostly looking for experience-based insights or reference points rather than strict formulas.
Thanks in advance!
https://redd.it/1onlpxe
@r_devops
I’m trying to understand how to estimate VPS resource requirements for different kinds of websites — not just from theory, but based on real-world experience.
Are there any guidelines or rules of thumb you use (or a guide you’d recommend) for deciding how much CPU, RAM, and disk to allocate depending on things like:
* Average daily concurrent visitors
* Site complexity (static site → lightweight web app → high-load dynamic site)
* Whether a database is used and how large it is
* Whether caching or CDN layers are implemented
I know “it depends” — but I’d really like to hear from people who’ve done capacity planning for real sites:
What patterns or lessons did you learn?
* What setups worked well or didn’t?
* Any sample configurations you can share (e.g., “For a small Django app with \~10k daily visitors and caching, we used 2 vCPUs and 4 GB RAM with good performance.”)?
I’m mostly looking for experience-based insights or reference points rather than strict formulas.
Thanks in advance!
https://redd.it/1onlpxe
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Dudes, I'm scared, I know it's scam, but what if it is not? Have you ever received a mail like this before? and how its going?
Dudes, I'm just a hobbys ( If you look in my linkedin, you'll notice I'm not a programmer ), I've learned programming, algorithms, design patterns, all by myself. I also publish articles on my Medium blog, documenting the concepts I've learned from my reading and online study.
I recently discovered an email written in Chinese in my Gmail spam folder and have translated its contents using Google Translate.
---
Tonghuashun AIME Program Invitation
Hello,
I am XXXXX, HR from Hexin Tonghuashun (Stock Code: 300033). We noticed your excellent background in development on GitHub, which highly matches the requirements for our AI Engineering Development position. The specific focus areas include, but are not limited to, algorithm application, algorithm engineering, and large AI model development.
Core Advantages of the Position
- Salary benchmarked against top-tier tech companies.
- Listed company stock incentives provided through the AIME Talent Double Hundred Plan.
- Participation in the R&D of AI products with millions of users.
- Tech Stack: Engineering (Java/Web/C++/etc.) and cutting-edge algorithms (Large Models/NLP/AIGC/Robotics/Speech/etc.).
Company Profile
Zhejiang Hexin Tonghuashun Network Information Co., Ltd. (Tonghuashun), established in 1995 and listed on the Shenzhen Stock Exchange in 2009 (Stock Code: 300033), is the first listed company in China's internet financial information services industry. We currently have over 7,000 employees, with our headquarters located in the beautiful and livable city of Hangzhou.
As an internet financial information provider, Tonghuashun's main business is to offer software products and system maintenance services, financial data services, and intelligent promotion services to various institutions, and to provide financial information and investment analysis tools to individual investors. To date, Tonghuashun has nearly 600 million registered users and over ten million daily active users. We have established business cooperation with over 90% of domestic securities companies, with a strong "moat" business ensuring stable cash flow for the company.
Supported Business
Based on comprehensive AI capabilities such as large models, NLP, speech, graphics, image, and vision, we currently cover multiple 2B and 2C application scenarios. Our numerous products include the intelligent investment advisory robot AIME, intelligent service, data intelligence, smart healthcare, AIGC, and digital humans. Targeting various regions including China, Europe and the US, the Middle East, and Southeast Asia, we are progressively realizing the path of technology commercialization and product marketization. The AI team has accumulated over ten years of experience, with hundreds of large model application scenarios implemented, trillions of user financial dialogue data points, and a self-built cluster of thousands of cards for computing power. We are one of the first domestic enterprises to receive cybersecurity administration approval for financial large models.
I look forward to discussing this further with you! You can contact me via:
- WeChat/Phone: XXXX
- Email: XXXX
If you are interested, please feel free to contact me at any time, and I will arrange a detailed conversation for you.
Wishing you the best of business!
HR XXX | Zhejiang Hexin Tonghuashun Network Information Co., Ltd.
Company Website: https://www.10jqka.com.cn/
----
Ok, now, I dont know what to do. I know this could be spam, but what if doesn't, I mean, links look real.
here is my git if you're interested in what they've seen: https://github.com/EDBCREPO
Have you ever received a mail like this before?
----
EDIT: f@@k, the link and job looks real: https://campus.10jqka.com.cn/job/list?type=school
https://redd.it/1ony2tl
@r_devops
Dudes, I'm just a hobbys ( If you look in my linkedin, you'll notice I'm not a programmer ), I've learned programming, algorithms, design patterns, all by myself. I also publish articles on my Medium blog, documenting the concepts I've learned from my reading and online study.
I recently discovered an email written in Chinese in my Gmail spam folder and have translated its contents using Google Translate.
---
Tonghuashun AIME Program Invitation
Hello,
I am XXXXX, HR from Hexin Tonghuashun (Stock Code: 300033). We noticed your excellent background in development on GitHub, which highly matches the requirements for our AI Engineering Development position. The specific focus areas include, but are not limited to, algorithm application, algorithm engineering, and large AI model development.
Core Advantages of the Position
- Salary benchmarked against top-tier tech companies.
- Listed company stock incentives provided through the AIME Talent Double Hundred Plan.
- Participation in the R&D of AI products with millions of users.
- Tech Stack: Engineering (Java/Web/C++/etc.) and cutting-edge algorithms (Large Models/NLP/AIGC/Robotics/Speech/etc.).
Company Profile
Zhejiang Hexin Tonghuashun Network Information Co., Ltd. (Tonghuashun), established in 1995 and listed on the Shenzhen Stock Exchange in 2009 (Stock Code: 300033), is the first listed company in China's internet financial information services industry. We currently have over 7,000 employees, with our headquarters located in the beautiful and livable city of Hangzhou.
As an internet financial information provider, Tonghuashun's main business is to offer software products and system maintenance services, financial data services, and intelligent promotion services to various institutions, and to provide financial information and investment analysis tools to individual investors. To date, Tonghuashun has nearly 600 million registered users and over ten million daily active users. We have established business cooperation with over 90% of domestic securities companies, with a strong "moat" business ensuring stable cash flow for the company.
Supported Business
Based on comprehensive AI capabilities such as large models, NLP, speech, graphics, image, and vision, we currently cover multiple 2B and 2C application scenarios. Our numerous products include the intelligent investment advisory robot AIME, intelligent service, data intelligence, smart healthcare, AIGC, and digital humans. Targeting various regions including China, Europe and the US, the Middle East, and Southeast Asia, we are progressively realizing the path of technology commercialization and product marketization. The AI team has accumulated over ten years of experience, with hundreds of large model application scenarios implemented, trillions of user financial dialogue data points, and a self-built cluster of thousands of cards for computing power. We are one of the first domestic enterprises to receive cybersecurity administration approval for financial large models.
I look forward to discussing this further with you! You can contact me via:
- WeChat/Phone: XXXX
- Email: XXXX
If you are interested, please feel free to contact me at any time, and I will arrange a detailed conversation for you.
Wishing you the best of business!
HR XXX | Zhejiang Hexin Tonghuashun Network Information Co., Ltd.
Company Website: https://www.10jqka.com.cn/
----
Ok, now, I dont know what to do. I know this could be spam, but what if doesn't, I mean, links look real.
here is my git if you're interested in what they've seen: https://github.com/EDBCREPO
Have you ever received a mail like this before?
----
EDIT: f@@k, the link and job looks real: https://campus.10jqka.com.cn/job/list?type=school
https://redd.it/1ony2tl
@r_devops
同花顺财经
同花顺财经__让投资变得更简单
核新同花顺网络信息股份有限公司(同花顺)成立于1995年,是一家专业的互联网金融数据服务商,为您全方位提供财经资讯及全球金融市场行情,覆盖股票、基金、期货、外汇、债券、银行、黄金等多种面向个人和企业的服务。
How are you handling these AWS ECS (Fargate) issues? Planning to build an AI agent around this…
Hey Experts,
I’m exploring the idea of building an AI agent for AWS ECS (Fargate + EC2) that can help with some tricky debugging and reliability gaps — but before going too far, I’d love to hear how the community handles these today.
**Here are a few pain points I keep running into 👇**
* When a process slowly eats memory and crashes — and there’s no way to grab a heap/JVM dump *before* it dies.
* Tasks restart too fast to capture any “pre-mortem” evidence (logs, system state, etc.).
* Fargate tasks fill up ephemeral disk and just get killed, no cleanup or alert.
* Random DNS or network resolution failures that are impossible to trace because you can’t SSH in.
* A new deployment “passes health checks” but breaks runtime after a few minutes.
**I’m curious**
* Are you seeing these kinds of issues in your ECS setups?
* And if so, how are you handling them right now — noscripts, sidecars, observability tools, or just postmortems?
Would love to get insights from others who’ve wrestled with this in production. 🙏
https://redd.it/1onzb40
@r_devops
Hey Experts,
I’m exploring the idea of building an AI agent for AWS ECS (Fargate + EC2) that can help with some tricky debugging and reliability gaps — but before going too far, I’d love to hear how the community handles these today.
**Here are a few pain points I keep running into 👇**
* When a process slowly eats memory and crashes — and there’s no way to grab a heap/JVM dump *before* it dies.
* Tasks restart too fast to capture any “pre-mortem” evidence (logs, system state, etc.).
* Fargate tasks fill up ephemeral disk and just get killed, no cleanup or alert.
* Random DNS or network resolution failures that are impossible to trace because you can’t SSH in.
* A new deployment “passes health checks” but breaks runtime after a few minutes.
**I’m curious**
* Are you seeing these kinds of issues in your ECS setups?
* And if so, how are you handling them right now — noscripts, sidecars, observability tools, or just postmortems?
Would love to get insights from others who’ve wrestled with this in production. 🙏
https://redd.it/1onzb40
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Paid Study Help us improve Virtual Machine Tools – $150 for a 60-minute interview
We’re conducting a paid research study to learn more about how professionals create, manage, and provision virtual machines (VMs) at work. Our goal is to better understand your workflows and challenges so we can make VM tools more efficient and user-friendly.
Details:
\- Compensation: $150 USD for a 60-minute 1:1 conversation
\- Format: Online interview via Zoom or Teams
\- Who we’re looking for: Anyone who creates or uses virtual machines, at any experience level or for any type of application
\- Priority: Participants with a LinkedIn profile linked to our platform will be considered first
If you’re interested, please send me a message or comment below and I’ll share the next steps.
Your feedback will directly help improve the tools used by thousands of professionals worldwide.
https://redd.it/1oo1fe7
@r_devops
We’re conducting a paid research study to learn more about how professionals create, manage, and provision virtual machines (VMs) at work. Our goal is to better understand your workflows and challenges so we can make VM tools more efficient and user-friendly.
Details:
\- Compensation: $150 USD for a 60-minute 1:1 conversation
\- Format: Online interview via Zoom or Teams
\- Who we’re looking for: Anyone who creates or uses virtual machines, at any experience level or for any type of application
\- Priority: Participants with a LinkedIn profile linked to our platform will be considered first
If you’re interested, please send me a message or comment below and I’ll share the next steps.
Your feedback will directly help improve the tools used by thousands of professionals worldwide.
https://redd.it/1oo1fe7
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Tired of applying everywhere - Looking for Fresher DevOps / Cloud Support / Linux Opportunity
Hey everyone,
I’m a recent Computer Science graduate actively looking for fresher roles in DevOps, Cloud Support, or Linux.
I’ve applied to many companies and portals, but most either ask for experience or never respond — it’s been really tough finding that first break.
I’ve learned and practiced:
Linux
AWS (EC2, S3, IAM, Lambda basics)
Docker & Kubernetes
Git/GitHub
CI/CD concepts
I’m genuinely passionate about DevOps and Cloud, and I’m just looking for that first opportunity to prove myself.
Preferably looking for roles in Pune or remote.
If anyone here knows of openings or referrals, I’d really appreciate your help 🙏
Thanks a lot for reading and supporting freshers like me!
https://redd.it/1oo3d4e
@r_devops
Hey everyone,
I’m a recent Computer Science graduate actively looking for fresher roles in DevOps, Cloud Support, or Linux.
I’ve applied to many companies and portals, but most either ask for experience or never respond — it’s been really tough finding that first break.
I’ve learned and practiced:
Linux
AWS (EC2, S3, IAM, Lambda basics)
Docker & Kubernetes
Git/GitHub
CI/CD concepts
I’m genuinely passionate about DevOps and Cloud, and I’m just looking for that first opportunity to prove myself.
Preferably looking for roles in Pune or remote.
If anyone here knows of openings or referrals, I’d really appreciate your help 🙏
Thanks a lot for reading and supporting freshers like me!
https://redd.it/1oo3d4e
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
India's largest automaker Tata Motors showed how not to use AWS keys
guy found two exposed aws keys on public sites, which gave access to \~70tb of internal data - customer info, invoices, fleet tracking, you name it
they also had a decryptable aws key (encryption that did nothing), a backdoor in tableau where you could log in as anyone with no password, and an exposed api key that could mess with their test-drive fleet
cert-in tried to get tata to fix it, but it took months of back-and-forth before the keys were finally rotated
link: https://eaton-works.com/2025/10/28/tata-motors-hack/ and https://news.ycombinator.com/item?id=45741569
https://redd.it/1oo402w
@r_devops
guy found two exposed aws keys on public sites, which gave access to \~70tb of internal data - customer info, invoices, fleet tracking, you name it
they also had a decryptable aws key (encryption that did nothing), a backdoor in tableau where you could log in as anyone with no password, and an exposed api key that could mess with their test-drive fleet
cert-in tried to get tata to fix it, but it took months of back-and-forth before the keys were finally rotated
link: https://eaton-works.com/2025/10/28/tata-motors-hack/ and https://news.ycombinator.com/item?id=45741569
https://redd.it/1oo402w
@r_devops
Eaton-Works
Hacking India’s largest automaker: Tata Motors
Tata Motors gave away the keys to their infrastructure and customer data on their public websites.
Those of you who switched from DataDog to Google Observability - do you miss anything?
The company I work for is switching from DataDog to Google's own offering, mostly driven by cost reasons. At surface level the offering seems to be par - but I wonder if we will discover things missing after it's too late?
https://redd.it/1oo36h7
@r_devops
The company I work for is switching from DataDog to Google's own offering, mostly driven by cost reasons. At surface level the offering seems to be par - but I wonder if we will discover things missing after it's too late?
https://redd.it/1oo36h7
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Cache Poisoning: Making Your CDN Serve Malicious Content to Everyone 🗄️
https://instatunnel.my/blog/cache-poisoning-making-your-cdn-serve-malicious-content-to-everyone
https://redd.it/1oo6g66
@r_devops
https://instatunnel.my/blog/cache-poisoning-making-your-cdn-serve-malicious-content-to-everyone
https://redd.it/1oo6g66
@r_devops
InstaTunnel
Cache Poisoning & CDN Exploits: How Your Cache Serve Malware
Learn how cache poisoning attacks enable CDNs and web caches to serve malicious content to all users. Discover how to detect, prevent and protect your caching
data democratization aka automation and management of data platforms
Hi folks, Are you guys aware of any platforms that can help with management of a number of users on large datalakes, what i mean by this say u have a product like databricks and we want to "user-wise" manage how much access someone has, we wanna stream line this by maybe this flow , user raises a request somehwere -> automated noscript grants access -> access revoked automatically within a set time,
also log who had what access etc etc,
ofc a custom solution is possible but i was hoping for any opinions on if anything similar to this already exists.
Thanks for yuour time have agood one
https://redd.it/1oo9x3u
@r_devops
Hi folks, Are you guys aware of any platforms that can help with management of a number of users on large datalakes, what i mean by this say u have a product like databricks and we want to "user-wise" manage how much access someone has, we wanna stream line this by maybe this flow , user raises a request somehwere -> automated noscript grants access -> access revoked automatically within a set time,
also log who had what access etc etc,
ofc a custom solution is possible but i was hoping for any opinions on if anything similar to this already exists.
Thanks for yuour time have agood one
https://redd.it/1oo9x3u
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
Learning friend
Is anyone here willing to learn Devops with me? I am a beginner
https://redd.it/1oo9js7
@r_devops
Is anyone here willing to learn Devops with me? I am a beginner
https://redd.it/1oo9js7
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
EKS Node Resource Limits
I am currently undertaking the task of auditing EKS Node resource limits, comparing the limits to the requests and actual usage for around 40 applications. I have to pinpoint where resources are being wasted and propose changes to limits/requests for these nodes.
My question for you all is, what percentage above average Usage should I set the resource limits? I know we still need some wiggle room, but say that an application is using on average 531m of Memory, but the limit is at 1000m (1Gb). That limit obviously needs to come down, but where should it come down to? 600m I think would be too close. Is there a rule of thumb to go by here?
Likewise, the same service uses 10.1mcores of CPU on average, but the limit is set to 1core. I know CPU throttling won't bring down an application, but I'd like to keep wiggle room there to, I'm just not sure how close to bring the limit to the average usage. Any advice?
https://redd.it/1oo78yq
@r_devops
I am currently undertaking the task of auditing EKS Node resource limits, comparing the limits to the requests and actual usage for around 40 applications. I have to pinpoint where resources are being wasted and propose changes to limits/requests for these nodes.
My question for you all is, what percentage above average Usage should I set the resource limits? I know we still need some wiggle room, but say that an application is using on average 531m of Memory, but the limit is at 1000m (1Gb). That limit obviously needs to come down, but where should it come down to? 600m I think would be too close. Is there a rule of thumb to go by here?
Likewise, the same service uses 10.1mcores of CPU on average, but the limit is set to 1core. I know CPU throttling won't bring down an application, but I'd like to keep wiggle room there to, I'm just not sure how close to bring the limit to the average usage. Any advice?
https://redd.it/1oo78yq
@r_devops
Reddit
From the devops community on Reddit
Explore this post and more from the devops community
GitOps role composition pattern for deployments?
Is anyone utilizing or has anyone utilized a cluster role-based composition pattern for deployments? Any other patterns?
Currently spinning up ArgoCD for current org and looking at efficiently implementing this for scalability.
At my previous org, we wound up having things a bit scattered about with \~30 AppSets and 30 applications (separate from appsets, for individual clusters).
It was manageable as we didn't change things much but I could see running into scaling issues as far as effort/maintenance goes down the road.
I would appreciate getting a second set of eyes to see if this makes sense or if I'm going to run into issues I haven't thought of: https://github.com/SelfhostedPro/ArgoCD-Role-Composition
https://redd.it/1ooejsr
@r_devops
Is anyone utilizing or has anyone utilized a cluster role-based composition pattern for deployments? Any other patterns?
Currently spinning up ArgoCD for current org and looking at efficiently implementing this for scalability.
At my previous org, we wound up having things a bit scattered about with \~30 AppSets and 30 applications (separate from appsets, for individual clusters).
It was manageable as we didn't change things much but I could see running into scaling issues as far as effort/maintenance goes down the road.
I would appreciate getting a second set of eyes to see if this makes sense or if I'm going to run into issues I haven't thought of: https://github.com/SelfhostedPro/ArgoCD-Role-Composition
https://redd.it/1ooejsr
@r_devops
GitHub
GitHub - SelfhostedPro/ArgoCD-Role-Composition: Example Repository utilizing roles to deploy applications to argo-cd clusters
Example Repository utilizing roles to deploy applications to argo-cd clusters - SelfhostedPro/ArgoCD-Role-Composition
How a tiny DNS fault brought down AWS us-east-1 and what devops engineers can learn from it
When AWS us-east-1 went down due to a DynamoDB issue, it wasn’t really DynamoDB that failed , it was DNS. A small fault in AWS’s internal DNS system triggered a chain reaction that affected multiple services globally.
It was actually a race condition formed between various DNS enacters who were trying to modify route53
If you’re curious about how AWS’s internal DNS architecture (Enacter, Planner, etc.) actually works and why this fault propagated so widely, I broke it down in detail here:
Inside the AWS DynamoDB Outage: What Really Went Wrong in us-east-1 https://youtu.be/MyS17GWM3Dk
https://redd.it/1ooi45v
@r_devops
When AWS us-east-1 went down due to a DynamoDB issue, it wasn’t really DynamoDB that failed , it was DNS. A small fault in AWS’s internal DNS system triggered a chain reaction that affected multiple services globally.
It was actually a race condition formed between various DNS enacters who were trying to modify route53
If you’re curious about how AWS’s internal DNS architecture (Enacter, Planner, etc.) actually works and why this fault propagated so widely, I broke it down in detail here:
Inside the AWS DynamoDB Outage: What Really Went Wrong in us-east-1 https://youtu.be/MyS17GWM3Dk
https://redd.it/1ooi45v
@r_devops
YouTube
Inside the AWS DynamoDB Outage: What Really Went Wrong in us-east-1
In this video, we’ll break down one of the most talked-about AWS incidents — the DynamoDB outage in the us-east-1 region.
I’ll take you inside the event to explain how a small internal DNS issue escalated into a large-scale DynamoDB failure affecting multiple…
I’ll take you inside the event to explain how a small internal DNS issue escalated into a large-scale DynamoDB failure affecting multiple…