• 0 Posts
  • 39 Comments
Joined 1 year ago
cake
Cake day: July 8th, 2023

help-circle

  • kyub@discuss.tchncs.detoLinux@lemmy.mlIs Linux As Good As We Think It Is?
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    2
    ·
    7 days ago

    Windows will continue to get more and more user-hostile as time goes on, and they want everyone to have a subscription to Microsoft’s cloud services, so they can be in total control of what they deliver to the user and how the user is using their services/apps, and they also will be able to increase pricing regularly of course once the users are dependent enough (“got all my work-related data there, can’t just leave”).

    The next big step that will follow after the whole M365 and Azure will be that businesses can only deploy their Windows clients by using MS Intune, which means MS will deploy your organization’s Windows clients, not your organization. So they’re always shifting more and more control away from you and into MS’ hands. Privacy is always an obvious issue, at the very least since Nadella is CEO, but unfortunately the privacy-conscious people have kind of lost that war, because the common user (private AND business sector) doesn’t care at all, so we will have to wait and see how those things will turn out in the future, they will start caring once they are being billed more due to their openly known behavior (driving, health, eating/drinking, psychology, …) or once they are being legally threatened more (e.g. your vehicle automatically reports by itself when you’ve driven too fast, or some AI has concluded based on your gathered data that you’re likely to cause some kind of problem), or once they are rejected at or before job interviews because of leaked health data or just some (maybe wrong) AI-created prognosis of your health. So I think there will be a point when the common user will start caring, we just haven’t reached that point yet because while current data collection and profile building is problematic because it’s the stepping stone to more dystopian follow-ups, it alone is still too abstract of an issue for most people to care about it. Media is also partly to blame here when they do reviews or news about new devices and then just go like “great camera and display, MUST BUY” and never mention the absurd amount of telemetry data the device sends home. MS is also partnering with Palantir and OpenAI which will probably give them even more opportunities to automatically surveil every single one of their business and private sector users. I think M365 also already gives good analytics tools to business owners to monitor what their employees are doing, how much time they spend in each application, how “efficient” they are, things like that. Plus they have this whole person and object recognition stuff going on using “smart” cameras and some Azure service which analyzes the video material constantly. Where the employees (mostly workers in that case) are constantly surveilled and if anything abnormal happens then an automatic alert is sent, and things like that. Probably a lot of businesses will love that, and no one cares enough about the common worker’s rights. It can be sold as a security plus so it will be sold. So I think MS is heavily going into the direction of employee surveillance, since they are well-integrated into the business world anyway (especially small and medium businesses) and with Windows in particular I think they will move everything sloooowly into the cloud, maybe in 10-15 years you won’t have a “personal” computer anymore, you’re using Microsoft’s hardware and software directly from Microsoft’s servers and they will gain full, unlimited, 100% surveillance and control of every little detail you’re doing on your computer, because once you hand away that control, they can do literally anything behind your back and also never tell you about it. Most of the surveillance stuff going on all the time already is heavily shrouded in secrecy and as long as that’s the case there will be no justice system in the world being able to save you from it, because they’d first need concrete evidence. Guess why the western law enforcement and secret services hunted Snowden and Assange so heavily? Because they shone some light into what is otherwise a massive, constant cover-up that is also probably highly illegal in most countries. So it needs to be kept a secret. So the MS (and Apple, …) route stands for total dependence and total loss of control. They just have to move slowly enough for the common user not to notice. Boil the frog slowly. Make sure businesses can adapt. Make sure commercial software vendors can adapt. Then slowly direct the train into cloud-only territory where MS rules over and can log everything you do on the computer.

    Linux, on the other hand, stands for independence. It means you can pick and choose what components you want, run them whereever and however you want, build your own cloud, and so on. You can build your own distro or find one that fits your use case the most. You’re in a lot of control as the user or administrator and this will not change considering the nature of open source / free software. If the project turns to sh!t, you’re not forced to stick with it. You can fork it, develop an alternative. Or wait until someone else does. Or just write a patch that fixes the problematic behavior. This alone makes open source / free software inherently better than closed source where the users have no control over the project and always have to either use it as it is or stop using it altogether. There’s no middle ground, no fixes possible, no alternatives that can be made from the same code base because the code base is the developer’s secret. Also, open source software can be audited at will all the time. That alone makes it much more trustworthy. On the basis of trustworthiness and security alone, you should only use open source software. Linux on its own is “just” the kernel but it’s a very good kernel powering a ton of highly diverse array of systems out there, from embedded to supercomputer. I think the Linux kernel can’t be beaten and will become (or is already) the objective best operating system kernel there is out there. Now, as a desktop user, you don’t care that much about the kernel you just expect it to work in the background, and it does. What you care more is UI/UX, consistency and application/game compatibility. We can say the Linux desktop ecosystem is still lacking in that regard, always behind super polished and user-friendly coherent UIs coming from especially Apple in that regard (maybe also a little bit by Microsoft but coherent and beautiful UIs aren’t Microsoft’s strong point either, I think that crown goes to Apple). That said, Apple is very much alike Microsoft in that they have a fully locked-down ecosystem, so it’s similar to MS, maybe slightly less bad smelling still but it will probably also go in the same direction as MS does, just more slowly and with details being different. Apple’s products also appeal to a different kind of audience and businesses than MS’ products do. Apple is kind of smart in their marketing and general behavior that they always manage to kind of fly under the radar and dodge most of the shitstorms. Like they also violate the privacy of their users, but they do it slightly less than MS or Google do, so they’re less of a target and they even use that to claim they’re the privacy guys (in comparison), but they also aren’t. You still shouldn’t use Apple products/services. “Less bad than utterly terrible” doesn’t equal “good”. There’s a lot of room between that. Still, back to Linux. It’s also obviously a matter of quality code/projects and resources. Big projects like the Linux kernel itself or the major desktop environments or super important components like systemd or Mesa are well funded, have quality developers behind them and produce high quality output. Then you also have a lot of applications and components where just single community developers, not well funded at all, are hacking away in their free time, often delivering something usable but maybe less polished or less userfriendly or less good looking or maybe slightly more annoying to use but overall usable. Those applications/projects could use some help. Especially if they matter a lot on the desktop because there’s little to no alternative available. On the server side, Linux is well established, software for that scenario is plentiful and powerful. Compared to the desktop, it’s no wonder why it’s successful on servers. Yes, having corporations fund developers and in turn open source projects is important and the more that do it, the more successful those projects become. It’s no wonder that gaming for example took off so hugely after Valve poured resources and developers into every component related to it. Without that big push, it would have happened very slowly, if at all. So even the biggest corpo haters have to acknowledge that in capitalism, things can move very fast if enough money is being thrown at the problem, and very slowly if it isn’t. But the great thing about the Linux ecosystem is that almost everything is open source, so when you fund open source projects, you accelerate their growth and quality but these projects still can’t screw you over as a user, because once they do that, they can be forked and fixed. Proprietary closed-source software can always screw over the user, no one can prevent that, and it also has a tendency to do just that. In the open source software world, there are very few black sheep with anti-user features, invasive telemetry, things like that. In the corporate software world, it’s often the other way around.

    So by using Linux and (mostly) open source products, you as the user/admin remain in control, and it’s rare that you get screwed over. If you use proprietary software from big tech (doesn’t even matter which country) you lose control over your computing, it’s highly likely that you get screwed over in various ways (with much more to come in the future) and you’re also trusting those companies by running their software and they’re not even showing the world what they put in their software.






  • Well, they’re only doing what they announced already like 1-2 years ago. So we knew it was coming. This is also accompanied by Google making YouTube more restrictive when viewed with adblockers. Google is (somewhat late, to be honest) showing its teeth against users who block ads. I always expected it to happen but it took them quite some time. Probably they wanted to play the good guys for long enough until most users are dependent on their services, and now their proprietary trap is very effective.

    On the desktop, you should switch to a good Firefox fork right now. Firefox can also be used but needs configuring before it’s good. The forks LibreWolf or Mullvad Browser are already very good out of the box. There’s the potential issue of the forks not being updated fast enough, but so far these two have been fast. Mullvad shares a lot of configuration with the Tor Browser, so using it may break some sites. LibreWolf might be “better” for the average user because of that, but otherwise I think Mullvad is the best Firefox fork overall.

    On mobile, Firefox-based browsers aren’t recommended, because on Android, the sandboxing mechanism of Firefox is inferior to that of the Chromium-based browsers. And on iOS, all browsers (have to) run on Apple’s proprietary Webkit engine anyway, but well this is Apple we’re talking about so of course it’s all locked-down and restricted. It’s one of the reasons I don’t even like talking about Apple that much, just be aware that as an iOS user, your choice doesn’t mean as much when it comes to browsers, and your browser might not behave like you think it does on other platforms.

    So on mobile, I’d suggest things like Brave, Cromite or Mull. Or Vanadium (GrapheneOS). If the browser doesn’t have built-in adblocking capability which sidesteps the MV3 restrictions, make sure to use an ad-blocking DNS server, so your browser doesn’t have to do it. But you still need it. Adblocking not only helps you retain your sanity when browsing the web in 2024, but it also proactively secures you against known and unknown security threats coming from ads. So adblocking is a security plus, a privacy plus, and a sanity plus. It’s absolutely mandatory. As long as the ad industry is as terrible as it is, you should continue using adblocks. All the time. On every device and on every browser.

    The ad industry is itself to blame for this. There could in theory be such a thing like acceptable ads, but that would require ads to be static images/text, not fed by personal data, and not dynamically generated by random scripts which could compromise your security, and not overly annoying. Since that is probably never going to happen, you should never give up using adblockers. Since they basically fight you by reducing your security and privacy, you have a right to defend yourself via technical means.


  • “We”, no. “Too many”, yes. In general, hard dependencies on proprietary software or services are often overlooked or ignored as potential future problems. Recent examples of this are Microsoft and VMware. Once the vendor changes things so that you don’t like anymore, or drives up prices like crazy, you’ll quickly realize that you have a problem you can’t solve other than switching, which you might not even be prepared to do short-term.

    The Windows world now experiences this because Microsoft is no longer interested in maintaining a somewhat quality operating system, they are mostly interested in milking their user base for data, and don’t hesitate to annoy or even disrupt their user base’s workflows in a try to achieve that goal.

    Many Windows users are currently looking at Linux because of this, but the more your whole workflow is based on dependencies to proprietary Windows-only software, the harder your time to switch will be. If you still use Windows today, you should at least start using more open source or cross platform software, which also will work on Linux, because you are on a sinking ship and there will probably be a time when you can’t take MS’ BS anymore and want to switch. Make it easier for you in the future by regarding Linux compatibility in the hard- and software you use today.


  • Well, Linux is like a juggernaut that’s inching ever closer in all sorts of areas (while already dominating in some areas). The time frame where it makes sense for Microsoft to spend increasing amounts of resources to maintain and further develop Windows is closing, and if you look closely, they’ve pretty much shown that Windows is not at all priority #1 anymore since at least Nadella became CEO. We also live in a world which is increasingly becoming OS agnostic, which is bad for Windows’ dominance and great for Linux, MacOS, and others (because there’s less and less relevant applications specifically requiring Windows). Of course, Linux on the desktop also grows stronger and more mature year after year, which further accelerates the change.

    There will also be some points in time which hugely accelerate things, like Valve going all-in on Steam Deck and Proton and to make Steam a more independent store/community platform, and also Microsoft making Windows worse and more user-hostile over time. From a business perspective, it makes sense for MS - they want to go full cloud (= full control), almost full removal of control for the user, and full ingestion of as much data from the user as they can - to sell it, utilize it for own purposes, and train AIs with it. It’s what increases profits in the short-term. A lot of companies are doing that kind of stuff. MS is just one of the more ruthless ones, which, again, makes sense, because they still have a big userbase to exploit. In the long-term, they’re damaging, no, DESTROYING Windows’ reputation as a half-decent OS (even among Windows fans) and driving more and more users to the alternatives. It’s kind of inevitable. MS’ striving for profit has doomed Windows, and soon, when no single company will be able to compete with the ever evolving Linux ecosystem anymore, Windows is also doomed. It’s kind of a law of nature now. It’s not a question of if, just when.

    (I’ve used both Windows and Linux extensively, Windows since MSDOS/Win3.x, Linux since 1998. About 10 years ago, I’ve switched exclusively to Linux and banned Windows into a VM only that gets booted less and less [I think it’s been off for 2 years already]). I, for one, welcome our new old Linux overlords.



  • Clickbaity titles on videos or news sites is the new standard. I watched it. The point he’s making is basically that music was harder to make/produce some 50 years ago, so there was more incentive to “make it worth the effort”, compared to today. And the 2nd point he makes is that music consumption is now so easy as well (listen to whatever you want instantly) compared to when you could only listen to something when you bought the physical album, that there’s also less incentive for the listener to really get involved into some albums.

    Personally I think these are valid points on the surface but they are not “the answer” to this kind of multi-faceted question. They’re at best a factor but we don’t know how big these factors are. Also I think one big reason he thinks that way is because he grew up in that environment and so he has a bias for “owning physical copies of albums”.

    I also think music hasn’t gotten worse, the market is just simply over-saturated because there’s just way too much music, you’ll never be able to listen to it all. And there are absolutely hidden gems or really good bands/artists forming even today, it’s just much harder to find them. Generally a problem of today’s age: it’s likely that what you’re looking for already exists, you just have to find it within a whole ocean of content.

    If you’re looking for innovative or non-standard stuff, you can always look at smaller artists or the indie scene, same is true for movies, games, music. The big producers always have a tendency to stick to what works and what’s proven to be popular so everything becomes similar. But smaller artists do not have to care about such things, they are ready to risk much more and in doing so, you might just create a real gem or something that was never or almost never tried before.


    • Closed source (has always been bad for an OS, a 1-US-company controlled blackbox at the heart of your “personal” computer)
    • Privacy nightmares (and getting worse)
    • Forced cloud integrations (and getting worse)
    • Forced AI integrations (and getting worse)
    • More bloat and ads (and getting worse)
    • More restrictions (e.g. local user accounts) (and getting worse)
    • More dark patterns to try to annoy the user and get him/her to accept something that MS wants (and getting worse)
    • More opt-out, on-by-default bad stuff being added (and getting worse)
    • There’s probably more…

    The question is wrong: it’s not why do you “still” hate Windows. I did like Windows 7. It was the last Windows I liked. After that, it’s just a downhill enshittification spiral. The only real question is: at which point will it be too oppressive for the common user that even the most common user will try to avoid it entirely. And I fear that there’s still more than enough room for MS to make Windows worse before enough people migrate away from it.



  • kyub@discuss.tchncs.detoArch Linux@lemmy.mlChoosing Next OS
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 months ago

    Arch is not at all problematic, however if you’re still inexperienced with Linux in general, there might be smaller issues with some packages which might be unsolvable or hard to solve in that particular case (but any experienced user can easily solve those by e.g. downgrading the problematic package until a fix is available, or by restoring a filesystem snapshot). My current Arch installation is almost 5 years old and I only had a couple of very minor individual package update issues, and one time where the system couldn’t boot anymore after an update, which could be desastrous for a newbie, but only for a newbie. So, any talk about Arch being unstable is most likely exaggerated. Windows 11 these days has more update failures than Arch, and Arch updates almost daily. Yes, Arch is not “perfectly stable” due to it being rolling release and receiving updates almost daily, but on the desktop or notebook that “less-than-perfect-stability” really doesn’t matter much unless you have some kind of allergy against breaking changes or spending 15-30min to fix something or get annoyed if you have to reboot. The fast updates and generally very up to date packages generally more than make up for the disadvantages. At least on the desktop and notebook. I’m not sure if I could recommend Arch on servers. Also, you should at least update Arch once a week (or more often). If you don’t update for multiple weeks, then updates might fail because Arch assumes that everyone is on or close to the most current updates. Or you might have to first update the pacman-keyring before updates work again. In any case, updating often is also a way to keep Arch more stable. If you don’t like updating often, don’t use Arch.


  • Not generally, however you might want to avoid any early dot-zero releases (e.g. 6.0.x). These tend to be a bit buggy with KDE Plasma, but the bugs get fixed soon. NVidia drivers should be better with the very latest updates, they are supposed to work well on Wayland now. But I don’t buy Nvidia, just AMD, so I’m not following this stuff closely.


  • Both are good. Librewolf is more like vanilla Firefox, just configured way better by default. Mullvad Browser is like a port of the Tor Browser (also based on Firefox) for the clear web (or for use with Mullvad’s VPN, or whatever). Also configured very well by default. Mullvad Browser has better anti fingerprinting stuff built-in but as a result of its unusual configuration some sites might be broken. Librewolf is kind of the opposite in that regard - sites won’t be broken but you’ll be easier to fingerprint. In any case, they both are at the top of the best Firefox variants I’d say.


  • kyub@discuss.tchncs.detoLinux@lemmy.mlDiscord rich presence on linux (game activity)
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    edit-2
    6 months ago

    Discord has a nice UI and lots of neat features, and it’s popular among gamers especially, but it can hardly be recommended. See https://www.messenger-matrix.de/messenger-matrix-en.html for a comparison with other communication programs. Yes, Discord has approximately the most red flags there can be. Discord is essentially spyware, it supports the least amount of encryption, security and privacy techniques out of them all, and everything you type, write, say and show on it is being processed and analyzed by the Discord server and probably in turn sold to 3rd parties. Discord can’t make a living from selling paid features only, they have to sell tons of user data, and since all data is basically unencrypted, everything’s free for the taking. Discord doesn’t even try to hide it in the terms of service or so. They just plainly state they’re collecting everything. Well, at least they’re honest about it. It’s a minor plus. If I had to use Discord, I’d only ever use the web browser version, and I’d at least block its API endpoints for collecting random telemetry and typing data (it doesn’t only collect what you sent, it also collects what you started typing).

    Matrix, on the other hand, is a protocol. Element is a well-known Matrix client implementing the protocol. On Matrix, everything is encrypted using quite state of the art encryption. It’s technologically much more advanced than Discord is. It’s also similar, but it won’t reach feature parity with Discord. Discord is a much faster moving target, and it’s much easier for the Discord devs because they need to, oh, take care of exactly nothing while developing it further. While adding a new feature to Matrix is much more complicated because almost everything has to be encrypted and still work for the users inside the chat channels.

    This is just broadly written for context. The two are similar, and you should prefer Matrix whenever possible, but I do get that Discord is popular and as is the case with popular social media or communication tools, at some point you have to bite the bullet when you don’t want to be left out of something. I’m just urging everyone to keep their communication and usage on Discord to an absolute minimum, never install any locally running software from them (maybe using sandboxing), and when you’re chatting or talking on Discord, try to restrict yourself to the topics at hand (probably gaming) and don’t discuss anything else there. Discord is, by all measurements I know, the worst privacy offender I can think about. Even worse than Facebook Messenger, WhatsApp and such stuff, because they at least have some form of data protection implemented, even if they also collect a lot of stuff, especially all metadata.



  • Choice of distro isn’t as important anymore as it used to be in the past. There’s containerization and distro-independent packaging like Flatpak or AppImage. Also, most somewhat popular distors can be made to run anything, even things packaged for other distros. Sure, you can make things easier for yourself choosing the right distro for the right use case, but that’s unfortunately a process you need to go through yourself.

    Generally, there’s 3 main “lines” of popular Linux distros: RedHat/SuSE (counting them together because they use the same packaging format RPM), Debian/Ubuntu, and Arch. Fedora and OpenSuSE are derived from RedHat and SuSE respectively, Ubuntu is derived from Debian but also stands on its own feet nowadays (although both will always be very similar), Mint and Pop!OS are both derived from Ubuntu so will always be similar to Ubuntu and Debian as well), and Endeavour is derived from Arch.

    I’d recommend using Fedora if you don’t like to tinker much, otherwise use Arch or Debian. You can’t go wrong with any of those three, they’ve been around forever and they are rock solid with either strong community backing or both strong community and company backing in the case of Fedora. Debian is, depending on edition, less up to date than the other two, but still a rock solid distro that can be made more current by using either the testing or unstable edition and/or by installing backports and community-made up to date packages. It’s more work to keep it updated of course. Don’t be misled by Debian’s labels - Debian testing at least is as stable as any other distro.

    Ubuntu is decent, just suffers from some questionable Canonical decisions which make it less popular among veterans. Still a great alternative to Debian, if you’re hesitant about Debian because of its software version issues, but still want something very much alike Debian. It’s more current than Debian, but not as current as a rolling or semi-rolling release distro such as Arch or Fedora.

    OpenSuSE is probably similar in spirit and background to Fedora, but less popular overall, and that’s a minus because you’ll find less distro-specific help for it then. Still maybe a “hidden gem” - whenever I read about it, it’s always positive.

    Endeavour is an alternative to Arch, if pure Arch is too “hard” or too much work. It’s probably the best “Easy Arch-based” distro out of all of them. Not counting some niche stuff like Arco etc.

    Mint is generally also very solid and very easy, like Ubuntu, but probably better. If you want to go the Ubuntu route but don’t like Ubuntu that much, check out Mint. It’s one of the best newbie-friendly distros because it’s very easy to use and has GUI programs for everything.

    Pop!OS is another Ubuntu/Mint-like alternative, very current as well.

    For gaming and new-ish hardware support, I’d say Arch, Fedora or Pop!OS (and more generally, rolling / semi-rolling release distros) are best suited.

    Well that’s about it for the most popular distros.


  • kyub@discuss.tchncs.detoLinux@lemmy.mlWhat is the /opt directory?
    link
    fedilink
    English
    arrow-up
    47
    ·
    edit-2
    6 months ago

    Let’s say you want to compile and install a program for yourself from its source code form. There’s generally a lot of choice here:

    You could (theoretically) use / as its installation prefix, meaning its binaries would then probably go underneath /bin, its libraries underneath /lib, its asset files underneath /share, and so on. But that would be terrible because it would go against all conventions. Conventions (FHS etc.) state that the more “important” a program is, the closer it should be to the root of the filesystem (“/”). Meaning, /bin would be reserved for core system utilities, not any graphical end user applications.

    You could also use /usr as installation prefix, in which case it would go into /usr/bin, /usr/lib, /usr/share, etc… but that’s also a terrible idea, because your package manager respectively the package maintainers of the packages you install from your distribution use that as their installation prefix. Everything underneath /usr (except /usr/local) is under the “administration” of your distro’s packages and package manager and so you should never put other stuff there.

    /usr/local is the exception. It’s where it’s safe to put any other stuff. Then there’s also /opt. Both are similar. Underneath /usr/local, a program would be traditionally split up based on file type - binaries would go into /usr/local/bin, etc. - everything’s split up. But as long as you made a package out of the installation, your package manager knows what files belong to this program, so not a big deal. It would be a big deal if you installed it without a package manager though - then you’d probably be unable to find any of the installed files when you want to remove them. /opt is different in that regard - here, everything is underneath /opt/<programname>/, so all files belonging to a program can easily be found. As a downside, you’d always have to add /opt/<programname>/ to your $PATH if you want to run the program’s executable directly from the commandline. So /opt behaves similar to C:\Program Files\ on Windows. The other locations are meant to be more Unix-style and split up each program’s files. But everything in the filesystem is a convention, not a hard and fast rule, you could always change everything. But it’s not recommended.

    Another option altogether is to just install it on a per-user basis into your $HOME somewhere, probably underneath ~/.local/ as an installation prefix. Then you’d have binaries in ~/.local/bin/ (which is also where I place any self-writtten scripts and small single scripts/executables), etc. Using a hidden directory like .local also means you won’t clutter your home directory visually so much. Also, ~/.local/share, ~/.local/state and so on are already defined by the XDG FreeDesktop standards anyway, so using ~/.local is a great idea for installing stuff for your user only.

    Hope that helps clear up some confusion. But it’s still confusing overall because the FHS is a historically grown standard and the Unix filesystem tree isn’t really 100% rational or well-thought out. It’s a historically grown thing. Modern Linux applications and packaging strategies do mitigate some of its problems and try to make things more consistent (e.g. by symlinking /bin to /usr/bin and so on), but there are still several issues left over. And then you have 3rd party applications installed via standalone scripts doing what they want anyway. It’s a bit messy but if you follow some basic conventions and sane advice then it’s only slightly messy. Always try to find and prefer packages built for your distribution for installing new software, or distro-independent packages like Flatpaks. Only as a last resort you should run “installer scripts” which do random things without your package manager knowing about anything they install. Such installer scripts are the usual reason why things become messy or even break. And if you build software yourself, always try to create a package out of it for your distribution, and then install that package using your package manager, so that your package manager knows about it and you can easily remove or update it later.