this post was submitted on 11 Jul 2023
520 points (99.8% liked)
Linux
48239 readers
503 users here now
From Wikipedia, the free encyclopedia
Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).
Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.
Rules
- Posts must be relevant to operating systems running the Linux kernel. GNU/Linux or otherwise.
- No misinformation
- No NSFW content
- No hate speech, bigotry, etc
Related Communities
Community icon by Alpár-Etele Méder, licensed under CC BY 3.0
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I'm loving the comments on the article.
I was previously unaware of the kernel doing such things.
Most of these are fringe cases nowadays, and often used in environments where the user has no control over the OS anyways. I don't really use Office at home (for the three times per year, LibreOffice is good enough and that's what most Windows users I know run at home anyways).
Also it's not as easy as to just "get Office, QuickBooks, Quicken, Adobe, and other major apps to run on Linux". The wine project is doing miraculous work already IMHO…
"Your article doesn't mention the real reasons, which conveniently enough I won't list either."
The reasons aren’t worth listing because they’re all known but here we go
You need to use linux shell to get anything done.
There, that’s the reason.
Linux will never be popular until you can do everything, and I mean, everything without entering a single command in a terminal.
You are overstating how much you need the terminal a bit. You can most certainly install and update software without the terminal. I get your point, but it's not 2006 anymore.
On which distro/s can I install all package types without opening terminal once?
Have you used Linux lately? You can do this in any distro with a modern desktop manager. Discover in KDE Plasma, Gnome Software, and similar in other desktop environments are installed by default in the DE and have been for like a decade.
You can install every native UI application and every Flatpak (or Snap) in every distro that ships with GNOME or KDE without opening terminal once. Not sure how the software center works for others but I'm sure they do the same.
Fedora, Ubuntu, Debian, Mint and many more. They all do it like this.
Need to install more than UI applications? Install dragora/Synaptic whatever GUI comes for your package manager. Not like you really need to do this because the average person only cares about the UI applications.
There are common programs you need to install via the terminal, you can’t even change sound playback quality without editing a conf file which requires sudo!
There is so much you need the shell for and until people stop defending it and start focusing on UX Linux will never be popular for your average user.
Out of interest, which programs do you need to install via terminal that concern the average user?
What do you consider changing "playback quality"?
Sampling rate? That can be changed in a config file without sudo (
~/.config/pipewire/pipewire.conf.d
), you shouldn't though because many applications expect 48000 as sampling rate. Unless you're doing studio recordings you want 48000.Correct, there is a lot of need for the shell, for power users. I don't really see anything that the average office and browser enjoyer needs to do in the terminal. You can even game now in most distros without opening the terminal once.
For example installing the GPU driver for an older GPU. Or installing the driver for an obscure printer, touchpad or other weird hardware.
Average user doesn't mean total noob. Installing Windows and the relevant drivers is something many users in the "Gamer class" can do. These guys usually don't to command line (except for maybe pinging something), but they are comfortable with installing and configuring stuff in GUI.
They understand how to google the driver to their weird hardware, download the .exe or .msi, start it and navigate the install wizard.
On Linux I've had it a few times that you e.g. have to unload/load kernel modules and stuff to get a driver working. I once even had it, that the Linux driver for a device was only supplied in source code to be compiled with an ancient version of GCC that wasn't available over the package manager. So then I spend an hour or two fixing compiler errors to upgrade that old source code to work with a current GCC.
Getting the same hardware to run under Windows meant downloading the .exe and running it.
And yeah, that's not something you'll do on a daily basis, but it is a huge roadblock for someone afraid of white text in a black window.
I feel like we can cherry pick situations on other operating systems where you might have to open a terminal window to solve an issue, but I agree that there are roadblocks that many won't even try to get past. There has been a load of progress around usability and informational resources for less advanced users over just the last few years. I feel the main barrier to entry is the willingness to learn something new.
I guess, many people here can't take it when people talk about the issues holding back Linux, considering the downvotes.
I think, if you like something, it is really important to talk about the issues it has, so that they can be improved. Blaming the users is not going to make more people switch.
These instances I posted weren't cherry picked. They where just what I encountered when setting up a single laptop.
I could tell you about the issues I had with my work laptop, where it was pretty difficult getting the VPN solution we use at work to run. We are using Teams and Outlook for work, both don't have official apps on Linux and the unofficial ones are really buggy.
Getting simple stuff like screen sharing to work under Wayland is basically impossible, which required me to revert to X11.
And sure, you can say that it's all edge cases, and people shouldn't be running Linux with a GPU, on an old device, with Microsoft tools or do screen sharing.
But if you say all of these common use cases are rare edge cases that shouldn't really be done on Linux, then you aren't talking about a general purpose desktop OS any more.
Older hardware and software that are made by companies who have hostile or ignorant stances towards FOSS are major contributors to at least some of the issues you mentioned. I can tell you that there is active development around solving some GPU/Wayland issues, but the limitations on what Linux can or can't do isn't fully the fault of Linux.
There is definitely room for improvement in Linux. The improvements in just the last three years shows that it has improved at a pretty brisk pace. Free and open community driven operating systems work toward the active needs of the community, so hopefully any issues or bugs you had got reported and you've actively checked up on them. I am making an assumption here, but if all of these issues you have had were extremely common, there would be every incentive for development of solutions to them.
There is some level of compromise that is needed when using proprietary software or hardware from hostile vendors or using some older hardware with Linux. This also goes for company supplied or required hardware/software. Linux might not be for everyone on every piece of hardware right now. The tradeoffs for having control over your hardware and software can sometimes be frustrating.
As for "blaming the users", I don't think I did that at all. I just feel like some folks prefer appliances over heavy machinery. That's personal preference. Sure, Linux should make onboarding as easy as possible, but in my opinion, the active pursuit of being #1 or #2 in desktop OS use is going down the wrong path. There is a certain type of person who chooses to go down the desktop Linux path and catering to their needs seems much more important for the long-term health of the OS.
I honestly don't see how Ubuntu installing the wrong driver is Nvidia's fault. Ubuntu has both the current and the legacy driver in their repos. When installing the OS it asks you whether you want to install proprietary drivers, which I agreed to, so it installed the non-legacy driver that does not work with the old GPU.
Also, when installing the correct driver on apt, there is no text prompt or CLI wizard (like in many other tools installed over apt) to actually load that driver into the kernel. I don't see how this is Nvidia's fault, as they haven't created the Ubuntu package.
That's squarely on the Ubuntu guys. That has literally nothing to do with hostile vendors.
I can't speak for Ubuntu or your situation as I don't have your issues and I don't use Ubuntu, but I would advise you reach out to the Ubuntu community with your issues and if you can't find a solution, file a bug report. They are a large community with a lot of engagement, so I would think that you might have luck either solving your problem or pointing their devs toward fixing the issue on their end. The squeaky wheel gets the grease.
Tbh, I spent enough time fixing it. I don't want to file bug reports and battle them though a bug log in my spare time. I do enough of that in my work time.
In my spare time I tend to use a computer to accomplish a task, not to fix the computer.
And that's where you hit the "average user" thing again. If an average user encounters a bug, they either live with the bug or give up. If you ask any of your non-dev friends how many of these has ever filed a bug report, I'd venture to say it's not going to be many.
And in the few cases where you actually see non-techy people filing a bug report, it's usually going to be on the level of "HELP! My PC is not working!" and the response will be "Closed for not following the template".
I don't know what to tell you. If you want to blame Ubuntu for your issues, but you aren't willing to go through the standard process of troubleshooting or filing a big report, maybe Linux isn't for you.
Dude, I am a dev, I am using Linux for over a decade. I spent hours fixing these issues. But we are talking about "The year of the Linux desktop" and Linux being used by people who aren't hardcore tech nerds or devs.
To quote from my edit that you probably didn't read yet:
The point I was trying to make is "Linux currently totally is not for everyone. There are a lot of issues that stop regular users from using Linux." and I was downvoted for that, and people kept argueing whose fault it is. And now you are coming to the conclusion that I was making in the beginning:
Maybe Linux isn't for everyone.
Because it currently totally isn't.
What really annoys me with this discussion (and I've had the exact same discussions hundreds of times) is that it always goes through the same cycle:
Combine that with the really misguided notion that anyone who isn't a tech crack must be dumb because otherwise they would be a tech crack, and it gets really dumb. People completely forget that there are lots of different topics where even most tech cracks have no clue about (e.g. medicine). I have ~20 years of experience as a dev. I know my way around the things I'm working with. I am the one everyone calls if they have issues with their computers. But I got no clue about fixing cars, medicine, or making clothes, just to name a few. And I am really grateful that if I go to my doctor, he treats me and doesn't tell me I'm dumb because I don't have a medical degree.
Okay buddy. You obviously didn't come here to do anything but grind your axe. 3% of desktop use is pretty cool, even if it's likely just a ton of Steam Decks. Anyway, have a nice time developing.
Well, fanboy, have fun pushing people out of Linux. People like you are a major part of the reason that it's still only 3%.
Just imagine that: There's a free OS that doesn't track you, doesn't serve you ads, you don't need the newest hardware to run it, it does almost anything that Windows does, and yet, it's got only 3 meagre percent of market share after 30 years.
You think that is because it's so easy to get into if you don't have a tech background?
Sadly, next to the technical hurdles and the bad UX, there is a really toxic community that is happy to shit on anyone who dares to say that Linux isn't perfect.
That's not quite my definition of "common".
The "Gamer class" is far from the average user, the average user doesn't even know what a GPU or a driver is and doesn't care. As long as the OS installs all drivers by default or the OEM has preinstalled them all is good.
Until there's no more drivers for that generation of GPU. The Windows 11 drivers for AMD only go down to the Vega 64, if you have a Fury X or a 7970 you're out of luck. Not that Windows 11 even lets you install on a machine that old.
AMDGPU goes down all the way to GCN 1.2, which means you can even run a 7970 on a modern Linux OS. Even out of the box if your distro has the legacy flags enabled.
It would be fantastic if there was more hardware that works out of the box in Linux, but that's up to the manufacturers. Until more people switch to Linux they don't bother and until they bother everybody complains that XY doesn't work on Linux.
As of right now the biggest hurdle is Nvidia without drivers included in Linux. Without a distro that takes care of installing their drivers they are essentially out of luck.
Using a GPU under Linux is not common? And installing Linux on old laptops isn't either?
I can't say anything about AMD, since the last time I had an AMD GPU is ~15 years ago.
When I installed an Ubuntu variant on my G580, which has a Geforce 635M it automatically installed the current driver for Geforce GPUs when I setup the OS, but that driver doesn't support the 635M. That one needs a legacy driver. And getting that to work was a major pain.
I first installed the legacy driver over apt, but it didn't do anything, because apparently installing the driver doesn't actually load the kernel module for the driver. So I had to load it manually, and it still didn't do anything. Turns out, uninstalling the original driver didn't unload it from the GPU either. So I had to re-install the old driver, unload the module, uninstall the old driver, install the legacy driver and load the legacy module. Took me a few hours to figure all of that out.
No way someone without CLI experience will be able to do that.
Installing drivers for an older GPU, obscure printer, touchpad or other weird hardware is not common.
Which is an issue with Nvidia, they have no drivers for that GPU for Windows 11 either. Not saying that this is not an issue but there is absolutely nothing Linux can do to make every legacy GPU work without help from Nvidia. It uses the open source driver out of the box, which works sometimes but not for everything and definitely not for gaming.
https://www.nvidia.com/en-us/drivers/results/180339/
Yes, they do.
Yes, they can. They literally have the correct (legacy) driver in the Ubuntu repo. But the autoinstaller installs the wrong driver during installing the OS. And if you try to manually install it, there is not even a text prompt in the CLI saying "You just installed that driver, do you want to actually use it to? (Y/n)".
They could have even gone so far as to make a CLI wizard (like many other packages do) or even a GUI wizard. But no, the package just installs and does nothing by default.
Also that is not correct. All the *buntu installers ask you when you install the OS whether you also want to have closed source drivers installed, and then it installs the closed source Nvidia drivers. Just the wrong ones.
That driver does not list the 635M but only the desktop version. Which is still impressive but I think they have separate drivers for their mobile chips? The latest driver listed for the 635M is only available for Windows 10 on their website.
Does
ubuntu-drivers devices
list the correct driver or is the recommended one too new? The driver packages in Ubuntu should install and activate themselves unless you have multiple installed, sounds like you ran into a bug.That does not change that Nouveau is used by default for the installer itself and by default for the OS if you don't select anything.
OS clouds and proprietary internets might shove another percent or 2 down the rabbit hole
As someone else has said, on distributions that go for ease of use, the terminal isn't really needed.
However, I do consider it a convenience feature even for users who are not savvy with it: You can either troubleshoot an issue by giving instructions like "Open application X, navigate to Option, open Tab, press Button, then enter Text, hit OK and repeat for each" or "copy and paste this command into your terminal". The amount of work on both sides is likely lower plus there's less room for error.
You would have to Give SUSE / OpenSUSE a try. It has Yast2-GUI so everything from setting up a samba share, ftp server, to kernal tweak, system services, and boot setup can be done entirely in the GUI environment. Very similar to how the older Windows Control Panel looked. Also One-click install for rpm files. Oh and system rollback if you blow up the system, no command line fixes needed.
Windows chance , and . depending on the language settings, so yeah so so simple and helpful :-/
While I do like Excel, its handling of values as dates is also a big issue that has hit a lot of people in the past – the format is just not very portable or exchangeable. It's not just an issue from Excel to other solutions… my point was rather that it's not a "Linux" issue and the way it was worded sounded like the kernel had something to do with it.
I get your point, but the guy you quoted also has a point. For a non-techy person it's really hard to understand the boundary of the OS, so where the OS ends and Apps begin. And tbh, even to a techy person, there isn't really a hard border there.
For example: Is the DWM part of the OS? On Windows, definitely. It's not the kernel, but it is the OS. You cannot remove or replace it.
On Linux, maybe, maybe not. It's definitely part of the Distro, but you can replace it. But on the other hand, on Linux you can even replace the kernel if you really want to. So maybe replaceability is not the criterium? But if it isn't, wouldn't that make everything that came in the distro part of the OS?
On Windows, that's kinda the case, with e.g. Edge being an integral, non-replaceable part of the OS.
And then you get into the territory of the "Linux is only the kernel" purists, that follow Stallmans fever dreams. They might say, Linux isn't actually an OS at all.
And at the latest once Stallman's speech has been quoted will anyone who is not a hardcore Linux philosopher say "Screw you guys, there is no point to this".
Nothing to do with the kernel, these are all application issues.