Rendered at 22:44:35 GMT+0000 (Coordinated Universal Time) with Cloudflare Workers.
guywithabike 1 days ago [-]
The worst part of all this is that GitHub's CTO and VP of Engineering sent out the usual "here's what we'll do to fix things" letter to their larger customers and, without exaggeration, it boiled down to: 1) "Here's a bunch of stuff we already did!" which... clearly isn't working, and 2) "We're continuing our Azure migration." also clearly not working.
So needless to say, if you depend on GitHub for critical business operations, you need to start thinking about what a world without GitHub looks like for your business and start working your way toward that. I know my confidence in GitHub's engineering leadership is at rock bottom.
Eji1700 1 days ago [-]
I could sorta see a situation where the reality is "we're in the middle of a miserable transition and it'll clean up when we're done" but I don't think anyone has confidence that's all it is at this point.
everforward 1 days ago [-]
Even that doesn’t really make sense to me, unless they’ve done it in a way where everything has to move at once.
Everywhere I’ve worked, if a migration is causing this much downtime then you kill the migration or slow it down. If every change has a 10% chance of bringing the site down, you only do a change every week or two until you can work out the kinks.
shrikant 1 days ago [-]
...or you keep fighting forward with the migration, because if it's seen as a failure then some pretty big heads will have to roll...
acedTrex 1 days ago [-]
I mean, they are seemingly breaking every week or two so that might be what they are doing.
suriya-ganesh 1 days ago [-]
also it should be noted that LinkedIn had a 5 year plan of migrating everything to azure but abandoned it after a year.
sysworld 1 days ago [-]
ooooh, they're migrating to Azure, now everything makes sense.
spondyl 1 days ago [-]
Here are some relevant excerpts from an October 2025 article[1]:
> In a message to GitHub’s staff, CTO Vladimir Fedorov notes that GitHub is constrained on capacity in its Virginia data center. “It’s existential for us to keep up with the demands of AI and Copilot, which are changing how people use GitHub,” he writes.
> The plan, he writes, is for GitHub to completely move out of its own data centers in 24 months. “This means we have 18 months to execute (with a 6 month buffer),” Fedorov’s memo says. He acknowledges that since any migration of this scope will have to run in parallel on both the new and old infrastructure for at least six months, the team realistically needs to get this work done in the next 12 months.
If you consider that six month parallel window to have started from the time of the October memo (written presumably at the start of October), then that puts us currently or past the point where they would have cut off their old DC and defaulted to Azure only.
Whether plans or timelines changed, I have no idea of course but the above does make for a convenient timeline that would explain the recent instability. Of course, it could also just be symptomatic of increased AI usage generally and the same problems might have surfaced at a software level regardless of whether they were in a DC or on Azure.
Putting that nuance aside, personally I like the idea that Azure is simply a giant pile of shit operated by a corporation with no taste.
>It’s existential for us to keep up with the demands of AI and Copilot
if by chance the CTO reads this, as a user of GitHub I would find it really existential if GitHub continues functioning as a reliable hub for git workflows (hence the name), and I have the strong suspicion nobody except for the shareholders gives a lick about copilot or 'AI' if it makes the core service the site was designed for unusable
ncruces 24 hours ago [-]
AI and Copilot increase the load on git workflows.
jwoq9118 1 days ago [-]
For GitHub to remain profitable they have to appease those shareholders you mentioned.
denkmoon 24 hours ago [-]
Why? What is the correlation between profit and shareholder sentiment (besides the fact that shareholders want said profits)? They don't really influence the operation of the business meaningfully.
kevin_thibedeau 22 hours ago [-]
Growth chart gotta go up. Only chumps run a business that makes a steady return.
denkmoon 20 hours ago [-]
Sure, but I think it's the wrong way around. Appeasing shareholders doesn't make you profitable, being profitable appeases shareholders. I think there is a wealth of evidence that appeasing shareholders actually impedes profits overall.
conception 24 hours ago [-]
Incorrect. They need to appease/trick/threaten/etc those that are paying for their services. Shareholders just demand they do so at the greatest (often short term) rate.
comice 1 days ago [-]
yeah currently working with Azure. what a PITA.
I wonder if the extended downtime is just due to the on-call engineers waiting for their azure auth tokens to refresh within azure's own damn network.
pm90 1 days ago [-]
i heard that they asked LinkedIn to do this too and they either refused or their systems were too complex so they refused to. Maybe that explains why LI availability seems ok
kleene_op 1 days ago [-]
Azure, the color of BSOD
cyanydeez 1 days ago [-]
they're not just migrating to Azure, they're vibrating to Azure!
AustinDev 21 hours ago [-]
Vibe coding and it's consequences.
"The evidence is clear: Either you embrace AI, or get out of this career." -Github CEO
"Sooner than later, 80% of the code is going to be written by Copilot. And that doesn’t mean the developer is going to be replaced." -Github CEO
justinko 22 hours ago [-]
There’s plenty of alternatives, but people continue to stay. Therefore, it’s not as bad as you think it is.
packetlost 1 days ago [-]
I second this. I'm done.
ryukoposting 1 days ago [-]
Is "migrating to Azure" the new "migrating to SAP?"
trvz 1 days ago [-]
That’s not for to … SAP.
zombot 7 hours ago [-]
The worst part is that users still stay and take the abuse.
proc0 1 days ago [-]
It's starting to really look like the AI effect. It might be coincidence but I've noticed a lot more downtime and bad software lately. The last Nvidia drivers gave me a blue screen (last week or so), and speaking about Windows, I froze updates last year because it was clear they were introducing a bunch of issues with every update (not to mention unwanted features).
I like AI but actually not for coding because code quality is correlated to how well you understand the underlying systems you're building on, and AI is not really reasoning on this level at all. It's clearly synthesizing training data and it's useful in limited ways.
davebranton 23 hours ago [-]
Interesting how many people "Like AI" because it's good at all the jobs other than the one they happen to make a living doing.
Did you hear about the screenwriters school in which the professors said to avoid AI for writing, but it's great for storyboards. And the storyboard school where the professors said the opposite?
The reality is that AI isn't actually "good" at anything. It produces passable ersatz facsimiles of work that can fool those not skilled in the art. The second reality of AI is that everyone is busy cramming it into their products at the expense of what their products are actually useful for.
Once people realise (1), and stop doing (2), the tech industry has a chance of recovering.
proc0 22 hours ago [-]
Yeah, I think I heard about that. Within certain domains it is certainly a useful tool. I would say things like online search are much nicer now (in that asking an AI is equivalent to searching online but it summarizes it for you). Online search fits the strengths of LLMs nicely, but right now it's being sold as a silver bullet, which it's not.
alsetmusic 3 hours ago [-]
I have no design talent but wanted to help my partner with some charts. She was making them in Excel. I had Claude Code build them as web docs and they look quite good. Probably had to give it around thirty instructions about changes, which was pretty inefficient, but then again I couldn't have created them myself and they look far nicer than the charts she got from Excel.
It's really just about recognizing what they can do well and applying them in the right moments.
danillonunes 22 hours ago [-]
It's a version of the Gell-Mann Amnesia effect.
someperson 1 days ago [-]
GitHub has been unreliable since before AI. Though it's definitely gotten far worse.
Seemingly the decline started with the Microsoft acquisition in 2018, and subsequent "unlimited private repository" change in 2019 (to match Gitlab's popular offer)
hparadiz 1 days ago [-]
One example is the search being broken for CI logs. It takes over your browser's search hotkey too. What happens is every stage of the log is collapsed so the search doesn't work until you trigger the expansion but if you attempt to search before expanding the search will never work after it's been initialized. It's pretty infuriating when you're trying to find something in a giant build log.
mghackerlady 15 hours ago [-]
The nvidia thing makes sense. If you get AI to write code for a platform you no longer really have an incentive to care that much about (windows) for a purpose you increasingly don't care about (actually drawing things to a screen), you're probably not going to test it as thoroughly as you used to
ivanjermakov 24 hours ago [-]
I bet on rushed Azure migration. A lot can go wrong it devops.
newbish 1 days ago [-]
I think maybe it's not that GitHub is using AI, but that the amount of AI slop going into GitHub may be more than they expected.
qbane 1 days ago [-]
Productivity is finite. If you pivot entirely to the AI stack, you're going to lose bandwidth for everything else. It's an opportunity cost problem.
holoduke 5 hours ago [-]
Indeed. I used to have 20. Now I have over 500 repos. All ai
thewhitetulip 21 hours ago [-]
We had more and more outages on EVERYTHING since AI. Not sure what they're doing.
boxingdog 1 days ago [-]
[dead]
paxys 1 days ago [-]
Took a full 8 years for a Microsoft acquisition to go to shit, which is probably a record. Kudos to the Github team for holding out this long.
AndroTux 1 days ago [-]
I’m still baffled that Minecraft is doing so well, despite the whole Bedrock thing. At this point I think Microsoft just forgot that they bought Mojang.
mghackerlady 15 hours ago [-]
I think they largely let mojang do its own thing, occasionally forcing them to make some dumb change that usually stays exclusive to their "bedrock edition". The mojang people capitulate since the original version and the one they actually develop for is largely untouched by microsofts decision making since the backlash for dumb decisions would lose infinitely more money than if they just let it continue to be a cash cow
spauldo 1 days ago [-]
They'd lose a whole lot of users if they killed Java edition, since the modded community is so large. They'd quickly find one of the Minecraft clones reaching feature parity. And there's no good reason for it - it's not like Java is a threat anymore.
AndroTux 1 days ago [-]
Exactly. So why isn't Microsoft doing just that? Isn't that how Microsoft usually handles things? Just look at Xbox. They essentially screwed up everything they could and then some.
PaulKeeble 1 days ago [-]
Its had its fair share of outages and outrageous changes that overreach the bounds as well. Its more stable than github is but its had at least 2 sessions of downtime this year that I recall and they were both quite long (day length).
joshribakoff 22 hours ago [-]
They don’t enforce or even default to 2fa to change the account email. In addition, they have no process to get a human to reverse account takeovers. Just a web form that tells you to call a number that redirects you back to a web form
hedora 19 hours ago [-]
On the other hand, they aggressively log out legitimate users, and require the master Microsoft account password to log back in (because your kids need access to your one drive settings, etc).
cedws 1 days ago [-]
If I remember correctly UK players can no longer chat at all until they verify their ID.
hedora 19 hours ago [-]
They've made it borderline impossible to log in, especially for children.
I just use an offline server, so I wouldn't notice if they had GitHub levels of availability.
7777332215 1 days ago [-]
Minecraft is a trick up their sleeve yet to be used. Manipulate and indoctrinate the youth.
Biganon 1 days ago [-]
Indoctrinated by cubic cows
merlindru 1 days ago [-]
GitHub has always been incredibly outage riddled no? This is not a MSFT thing
nine_k 1 days ago [-]
I don't remember that happening so much (if ever) in, say, 2016. But the frequency of noticeable incidents seemingly has been rising steadily since around 2023. The Azure migration apparently only exacerbated it.
gbear605 1 days ago [-]
Circa 2019, my office had a bell that we would ring whenever GitHub had an outage, and it was rung several times per week.
0x457 23 hours ago [-]
I remember seeing unicorn daily and "webhook delivery delayed" weekly. I think it got better, but also they got more traffic, now millions of agents read files separately over and over again.
IMO it's much better now.
georgel 1 days ago [-]
I remember it going down semi-regularly in the 2013+ era, and seeing HN posts about it. Especially if you were using a package manager reliant on GitHub like Cocoapods. It seems to me it is more "impactful" on the dev community now that they have gone past just being a centralized Git server for the team, to being the thing that does deploys and all sorts of other things.
bandrami 22 hours ago [-]
I think having a single 9 of uptime is a relatively new thing even for GH
theamk 20 hours ago [-]
It was not nearly as bad... I remember our company migrating to github.com, and believe it or not, it was significant performance/uptime benefit over our self-hosted instance.
(And the first thing to go was occasional 500's on github-hosted files.. the core service itself - git, PR, actions - were pretty stable until recently)
xeonmc 1 days ago [-]
How fast was Skype?
cm2187 1 days ago [-]
MSFT acquired Skype in 2011, so I would say only a few months:
To be fair a bunch of this is because the CEO after Nat Friedman (Thomas Dohmke) was pushed out in August 25.
carlosft 1 days ago [-]
And a ton of the top end ruby staff have left. Many of them ended up at shopify. There is a growing about of non ruby/rails code at github, but most of the system that people think of when they think github are ruby/rails.
ambicapter 23 hours ago [-]
Shopify is on the AI-everything train as well, we'll see how that goes.
workfromspace 1 days ago [-]
Who was also the last CEO, right? Is this a coincidence?
inaros 1 days ago [-]
Every day more Microsofty...they should rename to "Your Repository Needs To Restart To Apply Updates"
Waterluvian 1 days ago [-]
It's now safe to turn off your expectations.
kenhwang 1 days ago [-]
I wonder what the average career tenure of the userbase here is now, because Github was slow and flaky well before Microsoft got involved.
Maybe it wasn't as noticeable when Github had less features, but our CI runners and other automation using the API a decade ago always had weekly issues caused by Github being down/degraded.
jmtulloss 24 hours ago [-]
The best stretch Github ever had was post-acquisition when Nat Friedman as CEO.
bartread 1 days ago [-]
"It looks like you're trying to develop some software.
Would you like help?
- Get help with developing the software
- Just develop the software without help
[ ] Don't show me this tip again"
insin 17 hours ago [-]
The second and only other option not being "Maybe later"? Let me into the dream you're living in :)
hedora 19 hours ago [-]
If only their UI were that good!
morkalork 1 days ago [-]
Would you like to setups repository backups with OneDrive?
amarant 1 days ago [-]
Lol, someone should make a pre -commit hook that reboots your computer with a message like this!
corvad 1 days ago [-]
Just wait until github comes up with an outage tuesday.
mememememememo 1 days ago [-]
Down? No sir we are not down. There are elevated error rates and degraded performance.
karim79 1 days ago [-]
The update to .NET framework went badly and we need to reinstall Windows.
xtracto 1 days ago [-]
An isolated group of customers are experiencing elevated error rates and degraded performance.
FTFY.
(I've read AWS word it like that)
mememememememo 1 days ago [-]
Technucally a set is a subset of itself.
hirako2000 1 days ago [-]
I'm glad I moved over to forgejo. Being selfhosted, the UI loads faster. Most importantly, the thing is always responsive.
mfenniak 1 days ago [-]
As a developer working on Forgejo -- glad you like it!
hirako2000 1 days ago [-]
It hosts all the repositories backing applycreatures, we ran dozens of git projects on the same instance, have teams, you guys did a phenomenal work. I would say it's even easy to customise.
Please don't use that term; it makes them look bad :p /s
jeppester 1 days ago [-]
At this rate it will be a matter of time before a "Github is up" parody site reaches the top of HN
zelphirkalt 1 days ago [-]
Man, a while ago I thought: "It happens often, alright, but every 2 weeks? Sounds like a slight exaggeration." But it really is every 2 weeks, isn't it? If I imagine in a previous job anything production being down every 2 weeks ... phew, would have had to have a few hard talks and course corrections.
genewitch 1 days ago [-]
i once fixed a site going down several times a year with two t1.micro instances in the same region as the majority of traffic. Instantly solved the problem for what, $20/month?
Another site was constantly getting DDoS by Russians who were made we took down their scams on forums, that had to go through verisign back then, not sure who they're using now. They may have enough aggregate pipe it doesn't matter at this point
I once worked at a place with more micro services than engineers. We joked about "we have as many 8s of uptime as you need!"
0x3f 1 days ago [-]
> I once worked at a place with more micro services than engineers.
Currently consulting somwhere with 30 services per engineer. I cannot convince them this is hell. Maybe that makes it my personal hell.
KaiserPro 1 days ago [-]
"Its like family here!"
In that every night you're playing murder mystery, and its never fun.
0x3f 22 hours ago [-]
I would never trust my family with system design either.
NooneAtAll3 23 hours ago [-]
as a person that never touched webdev, I have a question
how is such service spam different from unix "small functions that do one thing only" culture?
why in unix case it is usually/historically seen as nice, while in web case it makes stuff worse?
0x3f 23 hours ago [-]
There are so many failures in microservices that just can't happen with a local binary. Inter-service communication over network is a big one with a failure rate orders of magnitude higher than running a binary on the same machine. Then you have to do deploys, monitoring, etc. across the whole platform.
You will basically need to employ solutions for problems only caused by your microservices arch. E.g. take reading the logs for a single request. In a monolith, just read the logs. For the many-service approach, you need to work out how you're going to correlate that request across them all.
Even the aforementioned network failures require a lot of design, and there's no standardization. Does the calling service retry? Does the callee have a durable queue and pick back up? What happens if a call/message gets 'too old'?
Also, from the other end, command line utils are typically made by entirely different people with entirely different philosophies/paradigms, so the encapsulation makes sense. That's not true when you're the one writing all the services, especially not at small-to-mid-size companies.
Plus, you already can do the single-concern thing in a monolith, just with modules/interfaces/etc.
msandford 1 days ago [-]
Oooof that's rough.
One strategy to convince is to get someone less technical than you to sit by you while you try and trace everything from one error'd HTTP request from start to finish to diagnose the problem. If they see it takes half a day to check every call to every internal endpoint to 100% satisfy a particular request sometimes that can help.
Also sometimes they just think "this is a bunch of nerd stuff, why are you involving me?!" So it's not foolproof.
0x3f 1 days ago [-]
Oh, my non-technical boss agrees with me already. It's actually the engineers who've convinced themselves it's a good setup. Nice guys but very unwilling to change. Seems they're quite happy to have become 'experts' in this mess over the last 5-10 years. Almost like they're in retirement mode.
The real solution is probably to leave, but the market sucks at the moment. At least AI makes the 10-repos-per-tiny-feature thing easier.
anotherjesse 1 days ago [-]
We pride ourself on 9 5s!
the_real_cher 1 days ago [-]
seven nines? That's nothing , bro we got twelve eights!
nuker 24 hours ago [-]
I have Royal flush :)
rdtsc 1 days ago [-]
From five nines to nine fives
Imustaskforhelp 1 days ago [-]
9% ? /s (though To be honest I genuinely wouldn't be surprised if things go down so bad too at this point either)
abound 1 days ago [-]
Unironically, I think 9% uptime would be "one-tenth of a nine".
brookst 1 days ago [-]
Are you saying 9.999% isn’t four nines?
munchler 1 days ago [-]
Can’t tell if this is intended as humor, but I LOL’ed.
the_real_cher 1 days ago [-]
It unarguably is.
mememememememo 17 hours ago [-]
Reliability where N is the number of nines:
1 - 10 ^ -N (multiply by 100 for percent)
So 9% is 0.09 for the calc
1 - 10 ^ -N = 0.09
So
10 ^ -N = 0.91
So
N = -log10 0.91
So 0.09 (9%) reliability is 0.0409586077 of a nine.
And running it thru... a tenth of a nine is 0.2056717653 or about 20.57% reliability
mememememememo 1 days ago [-]
90% would be one 9 following the sequence back.
99.99
99.90
99.00
90.00
sathish316 22 hours ago [-]
How much of this is due to Microsoft Culture of not innovating and buying leading companies with their revenue from windows/office and slowly destroying the aspects of companies that made them great in the first place?
Is all the recent GitHub downtime entirely attributable to GitHub AI Copilot related development? How hard can it be to reduce the blast radius of new AI features to not affect the core parts of hosting repositories? Because of Copilot everywhere, The UX has become bad and I had to click all over the place and on my profile to find repositories.
mayhemducks 1 days ago [-]
Does anyone else ever think "that code I just pushed into my repo just took down all of github..." whenever it goes down around the same time you sync your changes?
corvad 1 days ago [-]
And this is why I self host a lot of my Git stack with Gerrit...
mememememememo 1 days ago [-]
Or just make sure you git fetch repos into $other-place.
That helps with Git not so much issues etc.
corvad 1 days ago [-]
Yeah, I think especially Git mirrors can go a long too for maintaining availability and also for reducing load off main infra.
gchamonlive 1 days ago [-]
Just moved a project of mine to Gitlab. Created this very simple component with codex that will keep a mirror updated on GitHub for me, so I can focus development on Gitlab.
I do not care about much of it other than the git and API. I also sometimes use the Issues, although only with the API. But if it stops working sometimes, that is not too significantly a problem since the files can be sent after they start to work again; it does not have to be immediately.
MattIPv4 1 days ago [-]
Hitting 500s when trying to push branches and create PRs.
belter 1 days ago [-]
[dead]
sc__ 1 days ago [-]
Microslop
steeleduncan 1 days ago [-]
What has changed at GitHub to cause this?
smartmic 1 days ago [-]
AIpocalypse. Eaten too much Copilot dog food.
bartread 1 days ago [-]
Perhaps even AIslopalypse.
KaiserPro 1 days ago [-]
Looking at the status, its not one long outage, but lots of little ones, microslops if you will.
zahlman 1 days ago [-]
I've been using "slopocalypse". People already know AI is responsible, but slop existed before — e.g. conventionally generated SEO spam. It's just... so much worse now.
bartread 1 days ago [-]
"Slopocalypse": yeah, I like that. Easier to pronounce too.
At any rate, it seems like GitHub is back up now, so we'll see how long that lasts.
adzm 1 days ago [-]
Weird Al needs to capitalize on this whole AI/Al thing
pixelesque 1 days ago [-]
Possibly a combination of moving infrastructure to Azure, and also a significant increase in the number of PRs and commits due to Vibe-coding?
> GitHub has recently seen more outages, in part because its central data center in Virginia is indeed resource-constrained and running into scaling issues. AI agents are part of the problem here. But it’s our understanding that some GitHub employees are concerned about this migration because GitHub’s MySQL clusters, which form the backbone of the service and run on bare metal servers, won’t easily make the move to Azure and lead to even more outages going forward.
0xbadcafebee 23 hours ago [-]
Age-old lesson: change the tires on the moving vehicle that is your business when it's a Geo Metro, not when it's a freight train.
I'm sure the people with the purse strings didn't care, though, and just wanted to funnel the GH userbase into Azure until the wheels fell off, then write off the BU. Bought for $7.5B, it used to make $250M, but now makes $2B, so they could offload it make a profit. I wonder who'll buy it. Prob Google, Amazon, IBM, Oracle, or a hedge fund. They could choose not to sell it, but it'll end up a writeoff if the userbase jumps ship.
yoyohello13 1 days ago [-]
Vibe coding features.
staticassertion 1 days ago [-]
I assume this is all of the pains of going from "GHA is sorta kinda on Azure", which was a bad state, to "GHA is going full Azure", which is a painful state to get to but presumably simplifies things.
dec0dedab0de 1 days ago [-]
You never go full Azure
1 days ago [-]
qudat 1 days ago [-]
Their primary goal in the last year was to move to Azure. Any massive infra migration is going to cause issues.
seneca 1 days ago [-]
> Any massive infra migration is going to cause issues.
What? No, no it's not. The entire discipline of Infrastructure and Systems engineering are dedicated to doing these sorts of things. There are well-worn paths to making stable changes. I've done a dozen massive infrastructure migrations, some at companies bigger than Github, and I've never once come close to this sort of instability.
This is a botched infrastructure migration, onto a frankly inferior platform, not something that just happens to everyone.
the_real_cher 1 days ago [-]
A.I. but that acronym can mean a number of things.
Artificial intelligence, Azure integration, many other things.
paxys 1 days ago [-]
Senior engineers/leaders getting tired of Microsoft's shit and leaving.
rileymichael 1 days ago [-]
looking forward to the `addressing-githubs-recent-availability-issues-3` news post
overgard 1 days ago [-]
I remember back in the early Windows XP era when things got so bad that Microsoft basically had to make a hard pivot towards security and reliability.
I think they may need to do that once again. Almost every product of theirs feels like a dumpster fire. GitHub is down constantly, Windows 11 is a nightmare and instead of patching things they're adding stupid features nobody asked for. I think they need to stop and really look closely at what they're prioritizing.
tholford 1 days ago [-]
Setting up a Gitea instance is approachable, especially with agent assistance.
I guess vibe coding can't solve such problem for now...
ekropotin 1 days ago [-]
Remember when GitHub was cool? Pepperidge Farm remembers.
whalesalad 1 days ago [-]
I remember. My GitHub user ID is #5907, account created 2008-04-08T20:27:36Z. I think it is inevitable that all good things come to an end, but it's still a bummer to see.
Freedom2 1 days ago [-]
As do I. Mine is even earlier as well!
pylua 1 days ago [-]
Anyone else notice other Microsoft cloud services ( for instance inside azure ) with bad performance also?
I can’t be specific but we are constantly complaining.
keithnz 1 days ago [-]
No? Azures been rock solid for us.
pylua 1 days ago [-]
Front door did have a major outage last year.
sysworld 1 days ago [-]
The Azure management UI, yes, so slowww. But the services (VMs etc) have been good.
0xbadcafebee 24 hours ago [-]
I'm surprised nobody has tried to throw together a commercial alternative to GitHub. 50% of it is available as FOSS, the other 50% you can vibecode in a month (you can vibecode reliably, Microsoft/Google just suck at it). Afaict, reason we all keep using GitHub is it has a million features and isn't as ugly, difficult and slow as GitLab. (sorry GitLab, I love your handbook, hate your UX)
Would be easier for them to just tell us when it’s up these days
zombot 7 hours ago [-]
Good. Users should be abandoning this vibe-coded hellscape.
ygritte 7 hours ago [-]
You're underestimating Microsoft. They're fully capable of fucking up without the help of AI. Although I heard it's really good at that.
packetlost 1 days ago [-]
I've been sitting here waiting for a critical deploy to happen via GitHub Actions (I know, hour fault, we should have left ages ago). My patience for this bullshit is gone, I'm going to be pushing very hard to get us off of GitHub entirely except for public code mirrors going forward.
Edit: oh look, their site says all good, but I still have jobs stuck. What a pile of garbage.
I'm so sick of this.
s_u_d_o 1 days ago [-]
Can this downtime be quantified to actual monetary losses?
nasretdinov 1 days ago [-]
Must be Tuesday then
1 days ago [-]
butterlesstoast 1 days ago [-]
That Go rewrite continues to rear its head, eh?
justinholmes 1 days ago [-]
Moved to self hosted crowci instead of actions.
rco8786 23 hours ago [-]
GH is ripe for a disruptor right now.
newbish 1 days ago [-]
So am I the only one thinking that maybe GitHub is succumbing to the weight of AI slop that's coming in from all the vibecoding, clawbots, and other AI workflows?
esafak 1 days ago [-]
Microsoft products are so human, they stop working weekly as if they're observing some sort of sabbath ...
jiveturkey 19 hours ago [-]
I sure hope they created a restore point first.
lousken 1 days ago [-]
Github CEO must be on HN, right? If so, any comments?
They have not even bothered to implement entra login when they have their competitors login for years, do they even know what their product is? Or are you just a middle man for slop?
nomilk 23 hours ago [-]
Why don't companies with chronic outages mimic their stack from top to bottom (i.e. starting with a new domain), then before making a change, make the change on the duplicate stack and blast it with mock requests.
Might catch 90% of problems before they make it into the real stack?
E.g. every step of GitHub's migration to Azure could be mimicked on the duplicate stack before it's implemented on the primary stack. Is this just considered too much work? (I doubt cost would be the issue, because even if it costs millions, it would pay for itself in reduced reputational damage from outages).
EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
hedora 19 hours ago [-]
Downvotes are probably because that is what companies without chronic outages do.
If you'd ever worked on a codebase as terrible as I imagine GH's internals are and looked at the git history, you'd find two things:
1) fixing it would require rolling back 100's-1000's of engineer-years of idiocy that make things like testing or refactoring untenable
2) many prior engineers got part of the way through such improvements before leaving or being kicked out. Their efforts mostly just made it worse, because now you never know what sort of terribleness to expect when you open an unfamiliar file.
drewbug01 22 hours ago [-]
> EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
Because that's a monumental amount of work, and extraordinarily difficult to retrofit into a system that wasn't initially designed that way. Not to mention the unstated requirement of mirroring traffic to actually exercise that system (given the tendency of bugs to not show up until something actually uses the system).
nomilk 22 hours ago [-]
> that's a monumental amount of work
Agree, but look at the alternative; GitHub is constantly being savaged by users who (quite reasonably) expect uptime. Ignoring impacts on morale and reputation, damage to their bottom line alone might tens (hundreds?) of millions per year.
> mirroring traffic
yeah, I agree that's difficult, but it need to not be exact to still be useful.
worik 22 hours ago [-]
Testing? Who needs it when you have Copilot!
pbkompasz 1 days ago [-]
Vibe check?
cyanydeez 1 days ago [-]
That's just like your Vibe man; can you just copilot your wayout of these problems?
ransom1538 1 days ago [-]
Did MS finish the Hotmail transition?
duped 1 days ago [-]
Does github not do any kind of blue/green rollouts or what
GiorgioG 1 days ago [-]
I'm going to blame Claude Code!
19 hours ago [-]
rdedev 1 days ago [-]
My vscode slop session stopped in between. Maybe it's for the better
wrqvrwvq 23 hours ago [-]
"the food is terrible and such small portions!"
olivia-banks 1 days ago [-]
At least it happened after I did my work for the day... jfc!
messe 1 days ago [-]
For fucks sake.
I've been considering it for a while, but I'm definitely now pitching a move away from GitHub at our organization.
jbmilgrom 1 days ago [-]
h8 github so much. ahhhhhh
Imustaskforhelp 1 days ago [-]
You have GOT to be kidding me.
dylan604 1 days ago [-]
No, got to be kidding me day is next week.
Imustaskforhelp 6 hours ago [-]
Hey dylan, we got the got to be kidding me day today when claude showed outage (-:
rvz 1 days ago [-]
GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans maintaining it.
If there was a prediction market for when GitHub experiences an outage every week, then you would make a lot of money.
>GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans.
there are tens of thousands of stupid scripts hosted on github itself that have scheduled progmatic pushes or pulls to repos via cron jobs with millions and millions of users -- yeah LLMs accelerate the fire but let's not pretend that GH was some bastion of real-user-dom somehow at some point.
lesoutilsia 15 hours ago [-]
[dead]
corvad 1 days ago [-]
[dead]
darig 23 hours ago [-]
[dead]
3842056935870 1 days ago [-]
[dead]
unmayx 19 hours ago [-]
[dead]
bartread 1 days ago [-]
Fuck sake. Again?
Sorry, I realise this comment isn't up to HN's usual standards for thoughtfulness and it is perhaps a bit inflammatory but... look, I'd bet the majority of us on this site rely on GitHub and I can't be the only one becoming incredibly frustrated with its recent unreliability[0]?
(And, yes, I did enough basic data analysis to confirm that it IS indeed getting worse versus a year, two years, and three years ago, and is particularly bad since the start of this year.)
[0] EDIT: clearly not from looking at the rest of the comments in this discussion.
zahlman 1 days ago [-]
> I realise this comment isn't up to HN's usual standards for thoughtfulness
> And, yes, I did enough basic data analysis to confirm
Perhaps you'd consider showing us that analysis? That sounds like it would make a pretty substantive, thoughtful comment.
KaiserPro 1 days ago [-]
> consider showing us
Gaze upon the tapestry in which github paints it's failure with a thin copper red thread:
@KaiserPro has pasted the link to someone else's heatmap, which is really good. Mine was just an Excel spreadsheet with a graph that I'd intended to write a blog about but then got demotivated on because I was too busy with other things and I saw that heatmap as well. Maybe I will do a proper write up next time GitHub has an outage and I'm blocked by it.
So needless to say, if you depend on GitHub for critical business operations, you need to start thinking about what a world without GitHub looks like for your business and start working your way toward that. I know my confidence in GitHub's engineering leadership is at rock bottom.
Everywhere I’ve worked, if a migration is causing this much downtime then you kill the migration or slow it down. If every change has a 10% chance of bringing the site down, you only do a change every week or two until you can work out the kinks.
> In a message to GitHub’s staff, CTO Vladimir Fedorov notes that GitHub is constrained on capacity in its Virginia data center. “It’s existential for us to keep up with the demands of AI and Copilot, which are changing how people use GitHub,” he writes.
> The plan, he writes, is for GitHub to completely move out of its own data centers in 24 months. “This means we have 18 months to execute (with a 6 month buffer),” Fedorov’s memo says. He acknowledges that since any migration of this scope will have to run in parallel on both the new and old infrastructure for at least six months, the team realistically needs to get this work done in the next 12 months.
If you consider that six month parallel window to have started from the time of the October memo (written presumably at the start of October), then that puts us currently or past the point where they would have cut off their old DC and defaulted to Azure only.
Whether plans or timelines changed, I have no idea of course but the above does make for a convenient timeline that would explain the recent instability. Of course, it could also just be symptomatic of increased AI usage generally and the same problems might have surfaced at a software level regardless of whether they were in a DC or on Azure.
Putting that nuance aside, personally I like the idea that Azure is simply a giant pile of shit operated by a corporation with no taste.
[1]: https://thenewstack.io/github-will-prioritize-migrating-to-a...
if by chance the CTO reads this, as a user of GitHub I would find it really existential if GitHub continues functioning as a reliable hub for git workflows (hence the name), and I have the strong suspicion nobody except for the shareholders gives a lick about copilot or 'AI' if it makes the core service the site was designed for unusable
I wonder if the extended downtime is just due to the on-call engineers waiting for their azure auth tokens to refresh within azure's own damn network.
"The evidence is clear: Either you embrace AI, or get out of this career." -Github CEO
"Sooner than later, 80% of the code is going to be written by Copilot. And that doesn’t mean the developer is going to be replaced." -Github CEO
I like AI but actually not for coding because code quality is correlated to how well you understand the underlying systems you're building on, and AI is not really reasoning on this level at all. It's clearly synthesizing training data and it's useful in limited ways.
Did you hear about the screenwriters school in which the professors said to avoid AI for writing, but it's great for storyboards. And the storyboard school where the professors said the opposite?
The reality is that AI isn't actually "good" at anything. It produces passable ersatz facsimiles of work that can fool those not skilled in the art. The second reality of AI is that everyone is busy cramming it into their products at the expense of what their products are actually useful for.
Once people realise (1), and stop doing (2), the tech industry has a chance of recovering.
It's really just about recognizing what they can do well and applying them in the right moments.
Seemingly the decline started with the Microsoft acquisition in 2018, and subsequent "unlimited private repository" change in 2019 (to match Gitlab's popular offer)
I just use an offline server, so I wouldn't notice if they had GitHub levels of availability.
IMO it's much better now.
(And the first thing to go was occasional 500's on github-hosted files.. the core service itself - git, PR, actions - were pretty stable until recently)
https://trends.google.com/trends/explore?date=all&geo=GB&q=s...
Maybe it wasn't as noticeable when Github had less features, but our CI runners and other automation using the API a decade ago always had weekly issues caused by Github being down/degraded.
Would you like help?
- Get help with developing the software
- Just develop the software without help
[ ] Don't show me this tip again"
FTFY. (I've read AWS word it like that)
https://foja.applycreatures.com
Edit: it has a wonderful API so I posted the link it may tempt some to ditch MS/Azure hub.
Another site was constantly getting DDoS by Russians who were made we took down their scams on forums, that had to go through verisign back then, not sure who they're using now. They may have enough aggregate pipe it doesn't matter at this point
Currently consulting somwhere with 30 services per engineer. I cannot convince them this is hell. Maybe that makes it my personal hell.
In that every night you're playing murder mystery, and its never fun.
how is such service spam different from unix "small functions that do one thing only" culture?
why in unix case it is usually/historically seen as nice, while in web case it makes stuff worse?
You will basically need to employ solutions for problems only caused by your microservices arch. E.g. take reading the logs for a single request. In a monolith, just read the logs. For the many-service approach, you need to work out how you're going to correlate that request across them all.
Even the aforementioned network failures require a lot of design, and there's no standardization. Does the calling service retry? Does the callee have a durable queue and pick back up? What happens if a call/message gets 'too old'?
Also, from the other end, command line utils are typically made by entirely different people with entirely different philosophies/paradigms, so the encapsulation makes sense. That's not true when you're the one writing all the services, especially not at small-to-mid-size companies.
Plus, you already can do the single-concern thing in a monolith, just with modules/interfaces/etc.
One strategy to convince is to get someone less technical than you to sit by you while you try and trace everything from one error'd HTTP request from start to finish to diagnose the problem. If they see it takes half a day to check every call to every internal endpoint to 100% satisfy a particular request sometimes that can help.
Also sometimes they just think "this is a bunch of nerd stuff, why are you involving me?!" So it's not foolproof.
The real solution is probably to leave, but the market sucks at the moment. At least AI makes the 10-repos-per-tiny-feature thing easier.
1 - 10 ^ -N (multiply by 100 for percent)
So 9% is 0.09 for the calc
1 - 10 ^ -N = 0.09
So
10 ^ -N = 0.91
So
N = -log10 0.91
So 0.09 (9%) reliability is 0.0409586077 of a nine.
And running it thru... a tenth of a nine is 0.2056717653 or about 20.57% reliability
99.99
99.90
99.00
90.00
Is all the recent GitHub downtime entirely attributable to GitHub AI Copilot related development? How hard can it be to reduce the blast radius of new AI features to not affect the core parts of hosting repositories? Because of Copilot everywhere, The UX has become bad and I had to click all over the place and on my profile to find repositories.
That helps with Git not so much issues etc.
https://gitlab.com/gabriel.chamon/ci-components/-/tree/main/...
At any rate, it seems like GitHub is back up now, so we'll see how long that lasts.
https://www.forbes.com/sites/bernardmarr/2025/07/08/microsof...
To explain this one-word comment for those unfamiliar, see previously:
GitHub will prioritize migrating to Azure over feature development (5 months ago) https://news.ycombinator.com/item?id=45517173
In particular:
> GitHub has recently seen more outages, in part because its central data center in Virginia is indeed resource-constrained and running into scaling issues. AI agents are part of the problem here. But it’s our understanding that some GitHub employees are concerned about this migration because GitHub’s MySQL clusters, which form the backbone of the service and run on bare metal servers, won’t easily make the move to Azure and lead to even more outages going forward.
I'm sure the people with the purse strings didn't care, though, and just wanted to funnel the GH userbase into Azure until the wheels fell off, then write off the BU. Bought for $7.5B, it used to make $250M, but now makes $2B, so they could offload it make a profit. I wonder who'll buy it. Prob Google, Amazon, IBM, Oracle, or a hedge fund. They could choose not to sell it, but it'll end up a writeoff if the userbase jumps ship.
What? No, no it's not. The entire discipline of Infrastructure and Systems engineering are dedicated to doing these sorts of things. There are well-worn paths to making stable changes. I've done a dozen massive infrastructure migrations, some at companies bigger than Github, and I've never once come close to this sort of instability.
This is a botched infrastructure migration, onto a frankly inferior platform, not something that just happens to everyone.
Artificial intelligence, Azure integration, many other things.
I think they may need to do that once again. Almost every product of theirs feels like a dumpster fire. GitHub is down constantly, Windows 11 is a nightmare and instead of patching things they're adding stupid features nobody asked for. I think they need to stop and really look closely at what they're prioritizing.
https://about.gitea.com/
I can’t be specific but we are constantly complaining.
Edit: oh look, their site says all good, but I still have jobs stuck. What a pile of garbage.
I'm so sick of this.
They have not even bothered to implement entra login when they have their competitors login for years, do they even know what their product is? Or are you just a middle man for slop?
Might catch 90% of problems before they make it into the real stack?
E.g. every step of GitHub's migration to Azure could be mimicked on the duplicate stack before it's implemented on the primary stack. Is this just considered too much work? (I doubt cost would be the issue, because even if it costs millions, it would pay for itself in reduced reputational damage from outages).
EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
If you'd ever worked on a codebase as terrible as I imagine GH's internals are and looked at the git history, you'd find two things:
1) fixing it would require rolling back 100's-1000's of engineer-years of idiocy that make things like testing or refactoring untenable
2) many prior engineers got part of the way through such improvements before leaving or being kicked out. Their efforts mostly just made it worse, because now you never know what sort of terribleness to expect when you open an unfamiliar file.
Because that's a monumental amount of work, and extraordinarily difficult to retrofit into a system that wasn't initially designed that way. Not to mention the unstated requirement of mirroring traffic to actually exercise that system (given the tendency of bugs to not show up until something actually uses the system).
Agree, but look at the alternative; GitHub is constantly being savaged by users who (quite reasonably) expect uptime. Ignoring impacts on morale and reputation, damage to their bottom line alone might tens (hundreds?) of millions per year.
> mirroring traffic
yeah, I agree that's difficult, but it need to not be exact to still be useful.
I've been considering it for a while, but I'm definitely now pitching a move away from GitHub at our organization.
If there was a prediction market for when GitHub experiences an outage every week, then you would make a lot of money.
[0] https://news.ycombinator.com/item?id=47487881
there are tens of thousands of stupid scripts hosted on github itself that have scheduled progmatic pushes or pulls to repos via cron jobs with millions and millions of users -- yeah LLMs accelerate the fire but let's not pretend that GH was some bastion of real-user-dom somehow at some point.
Sorry, I realise this comment isn't up to HN's usual standards for thoughtfulness and it is perhaps a bit inflammatory but... look, I'd bet the majority of us on this site rely on GitHub and I can't be the only one becoming incredibly frustrated with its recent unreliability[0]?
(And, yes, I did enough basic data analysis to confirm that it IS indeed getting worse versus a year, two years, and three years ago, and is particularly bad since the start of this year.)
[0] EDIT: clearly not from looking at the rest of the comments in this discussion.
> And, yes, I did enough basic data analysis to confirm
Perhaps you'd consider showing us that analysis? That sounds like it would make a pretty substantive, thoughtful comment.
Gaze upon the tapestry in which github paints it's failure with a thin copper red thread:
https://www.githubstatus.com/