109
top 50 comments
sorted by: hot top controversial new old
[-] mrkite@programming.dev 83 points 1 year ago

I remember the 90s when both mac and windows crashed on a daily basis. When was the last time you saw a legitimate BSOD that didn't involve hardware failure? When was the last time you had to reset the PRAM on your mac just to get it to boot?

[-] Deely@programming.dev 18 points 1 year ago

Thats actyally very good point. Our phones x100 or x10K more powerful and complex than computers from 90s, but always works and very-very rarely need reboot.

[-] pkulak@beehaw.org 15 points 1 year ago

Kernels have gotten better. Professional tools have gotten better. Everything on Linux has gotten better. Compilers and drivers too.

Everything else is built by the lowest bidder and is absolute garbage. And unfortunately, it’s what most people interact with all day long.

[-] Noughmad@programming.dev 6 points 1 year ago

Eh.

This "everything else" are stuff that previously didn't even exist. There used to be only professional tools and a few games, now you have an app (or multiple apps) for everything.

And I'll take a garbage program over one that doesn't exist.

[-] nromdotcom@beehaw.org 3 points 1 year ago

There were lots of games back then. And many of them were as bad or worse than the shittiest shovelware and template swaps we've got today.

Thing is, most people don't remember the 200 Action Games 3 disc pack at the bottom of the bargain bin cause they sucked.

I'm not disputing that there is more "stuff" these days by raw numbers, with the barrier to creation and distribution of games and such dramatically lowered by ubiquitous and easy to use tooling. But I bet the ratios of good games to shitty games won't have changed too terribly much over the years.

load more comments (1 replies)
[-] JoeKrogan@lemmy.world 63 points 1 year ago* (last edited 1 year ago)

Not FOSS but commercial software is. The apps just get more bloated and want to suck even more data with each update. Then there is the sites that have hundreds of trackers and third party cookies from everywhere and need 1gb to display 🙄. OK maybe not 1gb but you get the gist.

[-] Dasnap@lemmy.world 31 points 1 year ago

Electron apps created with React can definitely push the boundaries of what 'acceptable' memory usage is.

[-] lysdexic@programming.dev 3 points 1 year ago

Electron apps created with React can definitely push the boundaries of what ‘acceptable’ memory usage is.

I have a pet theory that webview-based apps are popular only because currently there is absolutely no usable multiplatform desktop GUI framework. Therefore, developers have to resort to the one thing that works: load a webpage in a web browser.

Even React Native feels like a kludge in a way it converts React components to UI components.

[-] Draghetta@sh.itjust.works 16 points 1 year ago

Yep totally unheard of for foss software to get worse. Gnome 3 and kde 4 for example were universally acclaimed.

I think that the post author just neglects that software has become mindblowingly complex compared to the days of yore, if you put together all the features of netscape + win 3.11 + wordperfect + whatever other thing they were using in the 90s at any given point you don’t get 10% of the complexity of a contemporary productivity app (say outlook) let alone a full operating system.

It’s clear that the more complex something is the more things can break. It’s like complaining that F16s are worse than consumer 40€ drones because the former require maintenance every few hours of flight while the latter don’t.

[-] MagicShel@programming.dev 18 points 1 year ago* (last edited 1 year ago)

But if all you need is a drone and all anyone makes is an F-16, that is a shitty mismatch. I don't need an outlook that does all that shit, I just need to check my email, or at least set up a filter to send everything to the trash.

I don't need teams to do document management, I just need to chat with my team. I'll resend a document if it is needed for any reason. Companies are adding useless bloat to all of these things and then breaking the core functionality because they've made things hard. This is not progress.

Edit: fixed some iOS auto-incorrect. Apologies for any incoherence before.

[-] VonReposti@feddit.dk 7 points 1 year ago

It's kinda like having an F16 that are built to go through a McDrive. It's a feature, yes, but an utterly terrible one at that. Your job is to do one thing and do it well. I'll take the car for going to McDrive.

[-] hygieia@feddit.nl 6 points 1 year ago

He's not exactly comparing software to netscape or win 3.11 though, he's comparing version N of some software to version N-1 or N-2 and noticing that they're getting worse from release to release. Given the rate of new releases the complexity shouldn't be increasing that rapidly between releases so I'm not convinced that is the cause per se. I have to agree with the conclusion from the article, testing was more rigorous in the past than it is now. Both because there was less surface area to test back then and because time-to-market pressures were less due to the longer windows between releases.

load more comments (3 replies)
[-] Nioxic@lemmy.world 8 points 1 year ago

And of course, remember..

Early access / beta programs

Subscriptions for every fucking thing

[-] interolivary@beehaw.org 34 points 1 year ago* (last edited 1 year ago)

Everything is getting worse as companies are exclusively trying to squeeze more money out of everyone rather than build good products or services. Everything is done by fewer more overworked workers, with shittier components and features that are designed to extract money out of you rather than be useful or "good" (my favorite example is BMW's subscription based bench warmers.)

[-] Max_P@lemmy.max-p.me 8 points 1 year ago

That's really what's going on.

Back in the days, people took the time it was necessary to write the software. And managers trusted the engineers to say when it's ready or not.

Nowadays, the software world is managers going "yes we know the database's gonna blow up over the weekend without the query optimizations, but we want to build this new feature before the end of the week. We can deal with the database when it blows up over the weekend, that's why you guys are on-call."

I did not make this up, I've actually heard this. This is why modern software is so fucked up, not because we can't handle the complexity, because reliability and quality just isn't prioritized at all anymore. Gotta dish out new features every day and you're not allowed to work on fixing known critical bugs.

[-] interolivary@beehaw.org 5 points 1 year ago* (last edited 1 year ago)

I did not make this up, I’ve actually heard this

I was in the IT industry for about 20 years until I finally had enough a couple of years ago, so none of this is a shock to me 😅 "I've seen things you people wouldn't believe."

Smaller companies can still manake to make quality stuff more often than the giants, but the moment they actually make it into the bigger leagues (if they make it), they have to bring in a a lot of your average MBA types and business "intelligence" folks, and that's when things go down the shitter. They might be making more money but now everything is suddenly about "KPIs" - ie. usually badly defined and/or incorrectly calculated metrics about the service that supposedly reflect how good it is, but at best measure how addictive it is and how much money it makes you, but not how enjoyable it is or if it's actually even a good thing for your users. Not to mention how often those metrics are literal garbage based on such creative abuses of basic statistics that they aren't even wrong. It's astonishing how much trust people put in numbers some BI moron churned out, when the reality very often is that not only was the data collected wrong, but they made completely wrong conclusions about what their data represents ("I'll call the interval between these events the session length") and then applied some statistical methods on them that basically destroyed any information there could possibly have been (like taking the average of averages when the original populations have very different sizes). The CEO may not always understand the tech, but by gawd they do understand numbers, and all these KPIs just seem so convincing. When this line goes up it means that were doing good and on the right path, it's math right?

And now the C-suite is mostly made up of MBA types, the founders probably either left or were stuck in some dark corner where they won't bother anyone too much with their day drinking.

And now the internet service, or TV, or car, or washing machine, or whatever your company has been making is no longer really a service for the users, but more of a machine designed to bleed money out of them as efficiently as possible and that's run according to the tenets of what's essentially modern numerology in a suit and tie (and barely any better at describing anything you could call objective reality).

</rant>

[-] Max_P@lemmy.max-p.me 3 points 1 year ago

Yep, it's all about selling features now and not caring about the actual product itself at all anymore. Nearsightedness took over the whole industry: boost this quarter's metrics, heck, boost this week's numbers above anything else.

load more comments (1 replies)

I think that there are many potential causes, but I would like to add monopolization to the list.

Usually, a bad release spelled the demise of a company, because release times were so long that competitors could take advantage of a bad software release.

People aren't going to switch from windows because they release something bad or buggy, in that case it would already be dead. Windows isn't technically a monopoly, but they have a lot of inertia and there are many programs that only run on windows that people depend on. There is perhaps a limit to how bad windows can be before people abandon it en masse, but they can get away with a lot. The tech world is full of different companies and programs that are in monopolistic-ish positions.

[-] canis_majoris@lemmy.ca 7 points 1 year ago

The main issue is that Linux is too fragmented and absolutely not user friendly enough to be consumer-grade in most applications. Steam is doing their best with SteamOS and they have been making great strides in a lot of areas, and they've even allowed me to feel like I can run Linux as a primary OS without losing out on my main off-time workload of gaming. Stuff like DXVK and Proton have made amazing strides towards a gaming OS that isn't Windows.

Unfortunately too much shit goes wrong for the average user. Troubleshooting also becomes problematic when the community itself is fragmented on solutions. Often I will search up a problem and be recommended different solutions that are not using the tools I have available in favor of the other poster's favorite system. It's very annoying to say "I have problem X and have tools Y" and be told "Well, tool Y will do the job but tool Q will do it better".

I've been running Arch on a laptop recently and the first thing I had to do was troubleshoot networking. I looked at the router, wondered if I fucked up the config. Everything else connects fine, must be something else. Turns out that the clock was out of sync and it was preventing the OS from verifying any cryptography. The only time I've had that shit happen on Windows is on an old Surface RT that would randomly decide it was the year 3000.

Oh man, yeah I've been there with the arch networking issues. To be fair, I do think you sign up for some messing around when you decide to install arch, although it wouldn't hurt to make networking a bit easier. Troubleshooting on windows isn't very fun either, although you might not need to do it as often.

While it might be true that a lot of people are scared away by linux weirdness (or not, for all I know picking a beginner friendly distro and not doing anything weird might be a pretty decent experience, I've been using arch and doing weird stuff the past couple of years), most people don't even get far enough to install linux in the first place so the selection happens before that. Part of it is probably software compatibility, part of it might be that most people advocating linux are techy people who mostly talk about the techy reasons for why you should get linux, which aren't that appealing. Part of it could just be that people are resistant to change because it's annoying to have to learn a new system.

load more comments (2 replies)
[-] jadero@programming.dev 22 points 1 year ago

I have two hypotheses for why some kinds of software grow worse over time. They are not mutually exclusive and, in fact, may both be at work in some cases.

Software has transitioned from merely complex to chaotic. That is, there is so much going on within a piece of software and its interactions with other pieces of software, including the operating system itself, that the mathematics of chaos are often more applicable than logic. In a chaotic system, everything from seemingly trivial differences between two ostensibly identical chips to the order in which software is installed, updated, and executed has an effect on the operating environment, producing unpredictable outcomes. I started thinking about the systems I was using with this in mind sometime in the early 2000s.

The "masters" in the field are not paying enough attention to the "apprentices" and "journeymen. Put another way, there are too many programmers like me left unsupervised. I couldn't have had a successful career without tools like Visual Basic and Access, the masterful documentation and tutorials they came with, and the wisdom to make sure I was never in a position where my software might have more than a dozen users at a time at any one site. Now we have people who don't know enough to use one selection to limit the options for the next selection juggling different software and frameworks trying to work in teams to do the bidding of someone who can barely type. And the end result is supposed to be used by thousands of people on all manner of equipment and network connections.

One reason that open source software seems more reliable is that people like me, even if we think we can contribute, are mostly dissuaded by the very complexity of the process. The few of us who do navigate the system to make a contribution have our offerings carefully scrutinized before acceptance.

Another reason that open source software seems more reliable is that most of it is aimed at those with expertise or desiring expertise. At least in my experience, that cohort is much more tolerant of those things that more casual users find frustrating.

[-] lasagna@programming.dev 20 points 1 year ago

Is some software getting worse?

Fixed for ya.

[-] philm@programming.dev 14 points 1 year ago

Is most software getting worse?

Fixed for ya.

[-] lasagna@programming.dev 10 points 1 year ago

I wouldn't know. I only use some software.

[-] canpolat@programming.dev 20 points 1 year ago

I no longer look forward to updates.
[...]
It seems to me that some software is actually getting worse, and that this is a more recent trend.
[...]
Why does this happen? I don't know, but my own bias suggests that it's because there's less focus on regression testing. Many of the problems I see look like regression bugs to me. A good engineering team could have caught them with automated regression tests, but these days, it seems as though many teams rely on releasing often and then letting users do the testing.

The problem with that approach, however, is that if you don't have good automated tests, fixing one regression may resurrect another.

Every time I see a new update, I think: "I wonder what will break after this update" and postpone them as much as I can. Software updates shouldn't cause anxiety. But they do these days...

[-] huntrss@feddit.de 10 points 1 year ago

They used to cause anxiety in the past as well. But there was a window where - at least I - didn't fear them. Main reason why I still think they are necessary are security patches. But I do fear updates due to their tendency in breaking things.

[-] nebiros@programming.dev 17 points 1 year ago

everything went down when we allow web technologies be part of the desktop, everything electron or any other incarnation is an abomination

Amen. I miss actual desktop applications

load more comments (1 replies)
[-] Sekoia@lemmy.blahaj.zone 5 points 1 year ago

I can understand it tho. What other GUI toolkits have a markup language like HTML? It's so easy compared to having 50 billion different widgets, adding stuff through code is so verbose..

[-] JWBananas@kbin.social 5 points 1 year ago
[-] Sekoia@lemmy.blahaj.zone 3 points 1 year ago

Okay but do you expect me to actually know anything about what I'm saying?

(Jokes aside my bad TIL thanks)

[-] JWBananas@kbin.social 4 points 1 year ago

The real benefit with Electron is the whole write-once-run-everywhere goal that Java was supposed to originally achieve, combined with super fast prototyping.

Maybe one day we'll get a JIT/AOT version of HTML.

[-] balp@lemmy.world 15 points 1 year ago* (last edited 1 year ago)

It seems to me that the author doesn't remember all the struggles we had back then with bugs and features not working. And masses of needed functionality that never got skipped into the hands of users. It also strikes me that maybe there is a bit of nostalgia, just a bit of reluctance to change his ways. He found a workflow around the missing functionality that might be blocking for others and he has a harder time adjusting to the new functionality.

A bit like my father that refused to change his workflow, to make images for webpages (all static) he used for different Amiga programs because one could scale the images, one could edit them add lines and stuff, one for helping him make image maps, and they one so they could be converted to jpg/png as anim files used by everything else on the amiga didn't work well on the internet.

Bug testing back then was awful, we never had time to catch any issues but the biggest. The time plan for the release was fixed years ahead, the functionality that was needed was fixed years ahead. All the needed time for testing was eaten up by the developers working into the final skip to customers, trying to make the software actually run. It wasn't uncommon for test teams trying to cramp months of eating into a weekend to have the software skipped on Monday morning. Well including masses of needed bug fixes during that weekend that no one knew what code each issue was actually tested on. Remember that software version control system was almost not used, there was no CI build system all all software was built on some random developers workstation. Maybe, with some additional changes for his or her convenience. No software development has come a long way since the 90s. A very long way!

[-] demesisx@programming.dev 13 points 1 year ago

Yes. Case in point: there are at least 10 Lemmy iOS apps. I'll give you ten guesses on which ones are actually native Swift...

There are a quite a few Android apps in progress too. How many are written in Kotlin?

[-] aport@programming.dev 9 points 1 year ago

Jerboa is a native android app, so that's nice.

[-] philm@programming.dev 5 points 1 year ago

Yet it still feels sloppy (and I couldn't find why to the defense of Jerboa, I skimmed through the relevant code but couldn't find immediate issues and there's an open issue: https://github.com/dessalines/jerboa/issues/445)

I don't get why, with so much hardware power we still have these issues...

[-] philm@programming.dev 3 points 1 year ago

Is there really no native (SwiftUI) iOS app?

load more comments (4 replies)
[-] thingsiplay@kbin.social 12 points 1 year ago

@canpolat The article is written from the perspective of a Windows user. I'm not surprised. Software as a whole does not get worse; there are some software which get worse, and some technology getting in the way. And then there is software which get better, not worse too. If you cherry pick, then you can prove any point you want to proven right; especially on a very wide range of topics like software.

[-] pennomi@lemmy.world 9 points 1 year ago

Yeah this article is definitely suffering from a lot of biases in its logic. First off, they are cherry-picking, but there’s also survivorship bias in there.

For an obvious counter example, look at Blender. It’s gotten amazingly good even though it’s gotten more complex over the last 10 years or so.

[-] canpolat@programming.dev 7 points 1 year ago

I'm not sure it's that simple, really. And I definitely don't think this is limited to Windows. I agree with other comments that this is mostly related to complexity. The more complex the domain the more difficult it is to implement/maintain a good solution. Delivering the new shiny feature is more exciting for all people (product management, development, users, etc.) than to fix bugs. And if you don't have the resources/maturity to keep technical debt under control, the software quality will suffer over time. Free software may be the exception here as profit is not always the primary concern.

[-] thingsiplay@kbin.social 7 points 1 year ago

@canpolat My point was not being limited to Windows, but more that his view is limited to DOS/Windows world, but making general judgements about software. And because Windows and it's eco system of applications he listed gets worse, he extrapolates this to all software.

Let's look at Linux, which is probably the biggest software ever and used on every possible way one can imagine. It got better and better, even though it's extremely big and has a lot of complexity to it and does not want to break compatibility if possible. But I am not saying all software is like that. That's my point. Some software get better, some get worse.

[-] canpolat@programming.dev 6 points 1 year ago

And because Windows and it’s eco system of applications he listed gets worse, he extrapolates this to all software.

They admit that bias in the article:

[...] since I've always been working in the Microsoft tech stack, I use a lot of it. Thus, selection bias clearly is at work here.

Now, I mentioned free software as the exception. I don't have any data as to how big free software vs proprietary software. But I think his points extends at least to other proprietary software and is not limited to Windows.

Some software get better, some get worse.

I can agree with that.

load more comments (1 replies)
[-] Aux@lemmy.world 10 points 1 year ago

Modern software is getting more complex, yet the user base is getting less and less technically literate. At the same time software security is a bloody nightmare.

Back in the 90-s if your app crashed, it was an annoyance. Today if an app crashes, it's a possible attack vector to rob your users one way or another. You can't even create a simple and dumb web site anymore, you need a full security infrastructure around it, HTTPS, XSS protection, DDoS protection, etc, etc.

The rise of smart phones and internet in general, added billions of people to the pool of your users. Most of whom don't know the difference between address bar and Google. You have to invest a lot of resources into UI and UX as well as into protecting your application from people who don't understand how to operate computers.

Foolproof design, security and advanced feature sets make modern software ridiculously complex.

[-] mdhughes@lemmy.ml 7 points 1 year ago

Yes. At least since late '90s, and certainly the last 2 decades.

I blame the rise of frameworks, libraries, and IDEs. It's easier for someone who knows nothing to throw some software together and ship it. In the good old days, all software had to be written by someone who knew what they were doing, often in difficult tools. You had to think ahead and write code correctly, because you couldn't just ship patches every week.

And as junior devs get replaced by AI, there won't be any experience for any of them to learn how to do that.

[-] ExpensiveConstant@kbin.social 9 points 1 year ago

I blame the rise of frameworks, libraries, and IDEs. It’s easier for someone who knows nothing to throw some software together and ship it.

I very much disagree with this. Yes to an extent you don't need to know as much as you might have in the past but if we had to constantly reinvent the wheel, I don't think we would have nearly as many people entering/remaining in this field. Additionally well written frameworks and libraries can actually make your code safer since you don't have to reinvent the wheel and discover the pitfalls all over again. IDEs are also a net positive IMO. Errors next to the line of code that caused them, breakpoints, interactive debugging. These are all things I personally would find hard to live without. Necessities? Technically no. But good god do I not want to have to read build output unless necessary.

load more comments (2 replies)
[-] canpolat@programming.dev 8 points 1 year ago

In 2014 Robert Martin claimed that number of developers doubles every 5 years and says:

As long as that growth curve continues there will not be enough teachers, role models, and leaders. It means that most software teams will remain relatively unguided, unsupervised, and inexperienced. [...] It means that the industry as a whole will remain dominated by novices, and exist in a state of perpetual immaturity.

Not sure if the data can be confirmed or not, but if that's the case it will be difficult to maintain the best practices in our industry.

[-] balp@lemmy.world 4 points 1 year ago

I blame the rise of frameworks, libraries, and IDEs

Without good libraries and frameworks, we can hardly get any software working in today's environment. We get stuck with a slow development cycle and have software that doesn't do what the users want of it. A few years ago, I was at a customer using an old Linux distribution at their customer's site. For contractual reasons that was not upgrading to the latest version, they had skipped keeping up to date with changes as they came. Every step of development became a hassle and the good programmers there were not able to deliver features at any predictable rate. There were issues with HTTPS, most webservers of today mandate at least TLS1.2, but when the OS only supports SSLv2 and SSLv3, and TLS1.1, connecting to the internet, well gets hard.

Having to develop all functionality from the ground up, makes no features needed by the customers ever released. With most developers I have worked with using good libraries also makes the implementations less prone to have serious bugs in them.

[-] mark@programming.dev 3 points 1 year ago* (last edited 1 year ago)

I blame the rise of frameworks, libraries, and IDEs.

My thoughts exactly. Frameworks on top of frameworks with a lot of cruft that will incrementally make software slower and buggy.

That, coupled with the fact that business owners just want things shipped. Quality aside, I dont even think they care about products being good anymore 🥲

[-] Nommer@lemmy.world 2 points 1 year ago

What is a new person supposed to do to learn then? I've been stuck trying to figure out what the 5-8 different services, frameworks, libraries, and IDEs all do and I feel very overwhelmed.

load more comments (1 replies)
load more comments
view more: next ›
this post was submitted on 24 Jul 2023
109 points (91.6% liked)

Programming

17314 readers
138 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 2 years ago
MODERATORS