- GNOME removed all UI controls for setting solid color backgrounds, but still technically supports it if you manually set a bunch of config keys — which seem to randomly change between versions (see: https://www.tc3.dev/posts/2021-09-04-gnome-3-solid-color-bac...).
The pattern here seems pretty clear: a half-baked feature kept alive for niche users, rather than either properly supporting or cleanly deprecating it. Personally, I’d love to simply set an RGB value without needing to generate a custom image. But given the state of things, I’d rather have one solid, well-maintained wallpaper system than flaky background color logic that’s barely hanging on.
I checked in KDE, since I'm generally confused as to why it's not more popular now: in the wallpaper settings you choose `wallpaper type: plain color` and it gives you a color picker to set it.
It also shows you the screen you set it for, and a boolean to set it for all screens at once.
KDE used to be the "bloated" desktop way back when (I know, pretty silly and laughable now given the current state of things).
That cemented Gnome/Mate into a lot of major distros as the primary DME. Ubuntu being the most famous.
The QT licensing situation is also a bit of a bizarre quagmire. There are certainly people that don't like KDE for more ideological reasons.
Personally, none of this bothers me and it's what I use for my personal computer. KDE is just so close to exactly how I'm used to interacting with computers anyways growing up through the Win95 era. It is so close to the Windows experience you want to have.
Last time I used an Android (Galaxy) phone, to have a solid pitch-black background (which I thought made sense for modern phone displays energy-wise, besides looking pretty swell), I had to download — yes, D-O-W-N-L-O-A-D — some black image from some "Galaxy Store" thing or whatnot to achieve that. It was free, but it seemed like an exception there.
Something that should be a default option, or a single-tap switch in settings, turned into a chore consisting of a period of agonising disbelief, doubt, denial, search, and eventually bitter acceptance.
A black background on an AMOLED display (something Samsung Galaxy phones tended to have) would use less energy because “pitch black” on AMOLED is literally turning the underlying pixels off- with LED displays, that’s not possible.
> GNOME removed all UI controls for setting solid color backgrounds, but still technically supports it if you manually set a bunch of config keys — which seem to randomly change between versions
My analogous gnome experience was that on my tv-computer I was using 4x scaling, because TV and because my distance vision stinks.
At some point they decided 2x (3x?) scaling was enough for anyone and took away 4x, I didn't notice because I was already set at 4x and it continued working. Somewhat later they took away the backend, and then my system crashed with no error message immediately at login.
After much troubleshooting replaced a movie night, I inquired about the functionality being removed and was berated for using an undocumented/unsupported feature (because I was continuing to use it after the interface to set it had been removed, without my knowledge).
However, I've noticed, there's not much point in changing it. Showing the desktop is a waste of screen real estate because of the generations of abuse of desktop shortcuts. Even if you are careful, it becomes a cluttered wasteland (on Windows anyways). I just learned to never use the desktop for anything and always have windows up on various monitors.
My windows desktop remains pretty organised, occasionally there might be an app appear. My mac however I gave up on, it's just a mess of screenshots and files you have to drag n' drop from somewhere and the desktop is just where that ends up. I used to have a script that moved the screenshots but it's easier to just live in chaos.
I would love to store documents in the My Documents folder if applications actually had respect for me. Windows should never allow an application to just dump stuff in the Documents or Desktop folder without my permission.
I’m one of those people who can’t abide having desktop icons stick around. I don’t even use it as a staging area since I discovered Yoink. I kind of miss dragging disks to the trash to unmount them though. And the Oscar extension where he sang a little song every time.
I also ran into it with 15.4.0 and worked around it by creating an image that's the solid color I like to use. It turns out that the system-supplied solid color options are themselves just 128x128 pixel PNG images too:
The number of times my solid color preference has been replaced with all black over the last 10 years is absolutely astounding. I have no understanding of why this seems to be so difficult.
You know, I bought a MBP as a personal machine 6 months ago - my first Mac - and this post made me realize that I couldn't even tell you what its desktop looks like. I'm sure I saw it a few times when I was doing initial setup, but I don't remember anything about. I used to do things like set custom wallpapers, but probably around 10 years I pretty much stopped using the desktop entirely.
Maybe I'm still living in the 1990s, but isn't the graphics API call to paint a solid 2D rectangle infinitely faster than the call to blt a 2D image to the screen, with possible arbitrary scaling and clipping? Or is everything just natively textures now and we don't even have 2D drawing hardware?
When you say it that way, it's actually baffling that there's still separate code paths for solid background colors.
Just offer a background picker, and have it generate a 1x1 PNG, for the color selected. Just like that, you can use the image background code path for everything; and the code for generating the image almost certainly exists in the OS already. Maybe add some metadata to the generated image to filter it from the images picker, and you're done.
I recently acquired a ThinkStation P910 dual CPU Xeon E26xx with 64GB RAM and 1080GTX
Quite a capable machine for my uses.
Not supported in Windows 11. Maybe with some additional config? Can’t be bothered with something hat might turn out to be fragile and need more maintenance than I can be bothered with. That’s a young man’s gane.
Ok, I’m about due to give Linux another tickle anyways.
Hmm, which distro… can always give a few a spin.
Keep it simple, Pop!_OS.
Installed fast, no issues, runs fine, stable. Seems entirely usable.
I wouldn't go with Pop_OS with that hardware. The Nvidia GPU isn't supported by the new Nvidia driver and because System76 is hard at work writing Cosmic, their repositories are quite outdated. Support for things like Wayland is quite mediocre in the old drivers.
Switching to upstream (Ubuntu) with KDE would probably be more your speed.
As much as I'd like a machine like that, my 5 year old random Lenovo 10500 desktop is probably more useful as a daily driver machine than an older workstation class machine at the sacrifice of no ECC RAM. I bought it when it was 3 years old and will use it for 4 years then get rid of it before it hits the tail end, the power supply dies or something else goes wrong. You avoid all the weird problems, the depreciation, the energy costs running like that. And you gain things like relatively competent NVMe slots, USB-C and other luxuries. And the single core performance is better than Xeons of the era and earlier.
win11 ltsc works perfectly on it. With a solid background :D
It's funny to see this: after avoiding the Windows world for the last 25 years, back in the corporate world in the last few, I see this pattern with Microsoft tools all the time.
Teams not loading due to security issues, but notifications coming through with full content of messages. Ability to type a handful of words in cloud version of Word (or paste full documents) before security check catches up and requires me to set up a sensitivity label. Etc.
It mostly tells me about MS doing very bad software architecture for web apps, though apparently the desktop apps are not immune to it either.
It's not just MS. I think they might have fixed it now, but my personal favorite was when Google photos would send me a notification with a preview of an AI generated album of my photos they made for me even though the app did not now, nor ever have permissions (on Android) to look at said photos. And it too would then "catch up" and ask permission to see my files and I'd say "no" and then the preview would go away.
Similar with google docs, if you share a link to a doc, even if the doc is restricted access, anyone can see the thumbnail icon with the contents of page 1.
However when you're inside a note (which BTW, can also be converted into checkboxes, aka very simple TODOs), Google Keep, the note taking app from search giant Google, doesn't have search functionality for that specific note.
Besides the many small bugs, sometimes the missing functionality in Google apps is mind boggling.
That's odd, were you synchronizing your photos to Google Photos[1] in any way, from any device? Presumably they would've had to be synchronized to Google at some point for them to generate an album of said photos.
The team that wrote the preview portion of the app is a different team to the one that wrote the permission requesting part. They communicate asynchronously (as a team/org, but this probably is reflected in the app's architecture!), which means the outcome is eventually consistent! But you managed to observe one of those inconsistent cases!
I like to joke that Microsoft stuff is always 80%. Works very well for the obvious use cases, but then you are bound to run into weird issues as soon as you run into some edge case they haven't covered.
Makes me think it must have something to do with their corporate culture and how they work, since their developers, to my knowledge, never had a reputation for being weak. Maybe it's just because they have such a gigantic user base for everything they do, that 80% is the best business value. Though I doubt that a bit, as a third party developer, it makes me avoid their products where I can.
> Maybe it's just because they have such a gigantic user base for everything they do, that 80% is the best business value.
I was thinking about something similar recently. 80% of features take 20% of the time. For the hobby stuff I program, I want to make the best use of my time, so I skip the last 20% of features to make more room for stuff that matters.
I call it PDD: Pareto-Driven Development. Looks like you think Microsoft is similar.
I mean Microsoft has always had a reputation for shipping bad software, what that reflects on their developer teams is up to the audience. Personally I'd say it's reflective of a bad engineering culture that doesn't care about quality or craftsmanship due to lack of an incentive to ship good software but every incentive to ship software at scale, quality be damned.
Like in this article alone, that's a change that should never be made. It's an understandable bug but it's indicative of carelessness, both managerial (how does this hold up to code review?) and QA (why is no one asking why login splashes take 30 seconds?).
The most annoying one is that Windows machines have lost the ability for deep sleep. Laptops that slept perfectly 5 years ago are now left as 24/7 zombies, with the CPU, fans and hard disks running non stop.
I'm certain that some idiotic change just like the ones suggested in the article destroyed this perfectly working feature, and nobody is bothered to fix it because it would impact the latest harebrained scheme to make my 10 year laptop do AI in its sleep and suggest helpful ads about the things it couldn't help overhear while "sleeping".
I have never had a windows computer that was able to do any sort of sleep without crashing, either in the sleep mode or randomly a time after waking up. First thing I do is disable it, I have lost enough time trying to find the cause of it and the crashes usually stop if I don't use sleep modes.
Most of my computers and friends computers have been ASUS though, maybe that is a connection.
(Windows user since 3.11 but I don't think those had sleep modes :-)
They do have some sort of sleep but its very inefficient so Lenovo Thinkpads actually go into hibernation after an hour or so of sleep, to avoid the user waking up to an empty battery.
I just use hibernate instead. Marginally longer start up time and no more worries. For shorter periods I just leave it on (like when taking the laptop to a meeting room).
Windows kept waking up my old laptop at random times, draining the battery. Which lead to some embarrassment when I had replaced the default sounds with a comedy portal pack, and my backpack randomly yelled that the core temperature was critical when the battery nearly died
This relates to something that might have started around that time: The practice of displaying the splash screen for a finite time period, then showing the user environment before the software was fully started. It was suspected that both OS's and apps were doing this, because people notice when "the app takes too long to load."
Now you have to guess whether the software has really loaded or not before you start using it.
Or, in the corporate information systems world, we have to wait for the half-dozen security and monitoring systems to get done scanning memory and blocking while painstakingly logging to the clould. Only then do you get to watch a catatonic UI thread idly wait for every last piece of the underlying software to load.
Not only do they forbid devs to use Linux on their dev machines, they then proceed with cb.exe etc. Nothing shows how they value your time quiet as much as these artificial slowdowns they love to introduce. (Along with gigantic privacy issues, as it allows the employer to essentially look at a live feed of your desktop whenever they want)
I could understand it if your device needed special access (VPN to prod etc), but you usually can't do that either from the dev machines - and need to first connect to a virtual machine (via browser or rdp) to be able to do that...
It's clear to me that the intent with this 30-second timeout was that it's better for you to have a possibly half-broken but at least somehow usable desktop than be stuck with the loading screen forever, having to boot into a different OS to try to fix your main one.
> showing the user environment before the software was fully started.
and it has migrated to web apps today - where doing something causes the UI to show a loading/progress wheel, but it takes forever in actuality (or on start up of the webpage, you get a blank screen with placeholder bars/blurred color images etc).
I’m not sure if this was meant to be a pun, but “responsive design” has nothing to do with how quickly a UI loads. It’s about adapting to different screen sizes.
What would you suggest? Is it better to wait until the whole app is loaded to show anything? Or is the only solution to fix loading times in the first place?
Eh, in a case like this, without the 30 second “assume loaded” timeout, the system would be forever stuck in the loading screen for those impacted by the bugs. Sometimes it is better for your users to be optimistic that the system did indeed load.
> Also, I tend to stick with default configurations because it makes bug filing easier.
I've learned to use default configurations pretty much everywhere. It's far too much of a hassle to maintain customizations, so it's easiest to just not care. The exception is my ~50 lines of VS Code settings I have sync'd in a mysterious file somewhere that I've never seen, presumably on github's servers, but not anywhere I can see?
I only depends on an handful of tools (emacs, vim, lf, mpv, fish, foot,…) so I took the time to configure them and then just store the config in a git repo I sync everywhere. For personal computers I use stow. For remote machines, I just copy-paste. The nice thing is that those tools are so stable I could move to Debian stable and be OK.
> It's far too much of a hassle to maintain customizations
Is it? The vast majority of the time, I change settings/set things up the way I want, and then... leave them for literally years. Hell, I can directly restore a backup I have of Sublime Text from years ago and my customizations will work.
It is. I used to customize everything. On Windows 95/98/2000/XP - custom cursors, themes, icon packs, custom Windows loading screen, the works. When I used KDE (and Gnome for a while) and compiz came out, I enjoyed flaming windows. Same story - custom icon packs, make grub menu look nice, hell, custom kernels compiled for my CPU, etc.
Somewhere along the way I lost interest in customizing the OS. These days I routinely switch between MacOS, Windows and various Linux flavors on lots of computers. The only thing I may customize is I write my .vimrc from memory.
On my Android phones, I change the wallpaper and I disable animations. Otherwise, stock everything.
Now that I think about it, it can't be the time saved, surely I waste more time on HN. It likely correlates more with using computers for work as opposed to for fun and learning. Even the learning I do these days is rather stressful - if I can steal an hour or two on the weekend, I feel lucky, so spending time to customize the environment seems like a waste.
Maybe if life slows down, I'll find joy in customizing my OSes again.
Most of the time, yes. I maintained my vimrc for maybe 10-15 years before I gave up on it.
The hard part of maintaining a config is that there's no such thing as cost-free usage, it always takes a mental toll to change a config, to learn a new config, to remember which configs are in use and what they do, to backup configs, or at least to setup and maintain a config-auto-backup flow.
By far, the easiest mental model is just learning how everything works out of the box, and getting used to it. Then again, sometimes what people want is to do things the hard way for its own sake. That's probably part of why I kept going back to C over and over for so many years.
Reproducible self-contained configurations give most of the same benefit for bug filing.
Just your regular reminder that nix is good actually.
"I have a bug, you can get a full VM that reproduces it with 'nixos-rebuild build-vm --flake "github:user/repo#test-vm" && ./result/bin/run-*-vm'"
And the code producing that VM isn't just a binary blob that's a security nightmare, it's plain nix expressions anyone can read (basically json with functions).
And of course applying it to a new machine is a single command too.
(Would it be pedantic of me to say that I receive my fair share of bug reports on nix code I maintain, and when someone sends me their entire nixosConfig the very first thing I do is punt it back with a "can you please create a minimal reproducible configuration"? :D but your point stands. I think. I like to think.)
Agreed. I had a professor once who would say “The defaults were put there by people who probably know more about the software than you”. As long as you understand what the defaults are doing sometimes it’s more hassle messing with every option under the sun.
Same except my teacher’s version was “just hit next/yes” for every option when installing software, in an era before that’d get you Adobe Reader and McAfee malware.
These days the defaults are almost certainly oriented towards controlling or extracting the most value from you, be it invasive spyware, constant intrusions, or sub-optimal UI.
“Comfort food”. That’s so funny. I still use motif window manager with steelblue4 desktop and wheat xterm background since aix into Linux. That was my first default in 1989 college and nothing has improved since. (Gnome, kde and the like make me want to upchuck).
What's the biggest thing that's improved? We have 4x the pixels, so we spend 4x the rendering time to draw everything with 4x as many pixels, when it works, and complain when it doesn't.
Would have been easier to stick with the pixel density we had.
Oh, and we have to wait a frame to see everything because of compositing that I still don't quite understand what it's supposed to do? Something something backing store?
My "comfort food" in this article is the realization that no matter how big, how advanced a team can be -- we all make (and ship) really dumb changes to production. A bolted-on wrapper if() statement that spans a bit too far is classic.
Many, many years ago, when I still used Windows recreationally, I used to edit the value of a specific Windows registry key, replacing "explorer.exe" with "cmd.exe". This would prevent Windows from runningg explorer.exe and launching the "desktop" with wallpaper, icons, etc. This produced what seemed more like a UNIX window manager, a solid backgground, where each window is a Microsoft cmd.exe shell, a classic Windows black box with a blue bar across the top and thin grey borders. I could then launch applications, such as the ones in the C:\windows\system32 folder from a command prompt, e.g., taskmgr.exe. For me, this made Windows feel much faster and more robust that when using explorer.exe. Certainly it was lighter weight.
More recently, long after I stopped using Windows but still many years ago, I was reading an article about Arthur Whitney. It had a photo which seemd to be at home, maybe in a furnished garage, and in the background was a desktop computer running Windows. The only window open was a cmd.exe. I am not suggesting anything. It is just something I always remember.
Perusing some recent Microsoft documentation I noticed this:
There is/was a mode for this in Windows server. I also seem to remember a low cost/free windows for server or embedded meant for iot or whatever that featured cmd only without explorer.
To me, this falls in the category of bugs I think of as "systemic bugs" or "type bugs". If login components were passed a token, then you could make the token's destructor automatically flag that the process is done. Then this bug would be more-or-less impossible to write.
Because they made it a runtime thing - "components just have to remember to do this", the code structure itself affords this bug.
There was a similar bug at facebook years ago where the user's notification count would say you had notifications - and you click it, and there aren't any notifications. The count was updated by a different code path than the code which inserted notifications in the list, and they got out of sync. They changed the code so both the notification count & list were managed by the same part of the system, and the all instances of the bug went away forever.
For as long as I've used the official app on iOS (since they killed third party apps... which also got me using reddit SIGNIFICANTLY less) it's had the issue where it'll send a push notification for a message reply or whatever, you can click it, view the message in the thread, reply to it, whatever. The push notification will be gone from your iPhone, but it never actually clears the notification from within the app. So you'll still see the notification in your in-app inbox so you have to go in there manually and click it again.
A bit meta, but I've come to look forward to the "Why did <bizarre behaviour> happen with <windows version>" headlines that herald a Raymond Chen post. These are always fascinating.
Agreed. The Old New Thing is such a wonderful collection of Windows oddities. Its the only blog that has explained away computing myths surrounding Windows that I've held since childhood.
Some additional examples beyond the OP:
- In the latest macOS, trying to set a custom solid color background just gives you a blinding white screen (see: https://discussions.apple.com/thread/256029958?sortBy=rank).
- GNOME removed all UI controls for setting solid color backgrounds, but still technically supports it if you manually set a bunch of config keys — which seem to randomly change between versions (see: https://www.tc3.dev/posts/2021-09-04-gnome-3-solid-color-bac...).
The pattern here seems pretty clear: a half-baked feature kept alive for niche users, rather than either properly supporting or cleanly deprecating it. Personally, I’d love to simply set an RGB value without needing to generate a custom image. But given the state of things, I’d rather have one solid, well-maintained wallpaper system than flaky background color logic that’s barely hanging on.
It also shows you the screen you set it for, and a boolean to set it for all screens at once.
KDE used to be the "bloated" desktop way back when (I know, pretty silly and laughable now given the current state of things).
That cemented Gnome/Mate into a lot of major distros as the primary DME. Ubuntu being the most famous.
The QT licensing situation is also a bit of a bizarre quagmire. There are certainly people that don't like KDE for more ideological reasons.
Personally, none of this bothers me and it's what I use for my personal computer. KDE is just so close to exactly how I'm used to interacting with computers anyways growing up through the Win95 era. It is so close to the Windows experience you want to have.
Something that should be a default option, or a single-tap switch in settings, turned into a chore consisting of a period of agonising disbelief, doubt, denial, search, and eventually bitter acceptance.
There's the peak GNOME experience.
At some point they decided 2x (3x?) scaling was enough for anyone and took away 4x, I didn't notice because I was already set at 4x and it continued working. Somewhat later they took away the backend, and then my system crashed with no error message immediately at login.
After much troubleshooting replaced a movie night, I inquired about the functionality being removed and was berated for using an undocumented/unsupported feature (because I was continuing to use it after the interface to set it had been removed, without my knowledge).
I'll never use gnome again if I can help it.
I suspect this is related to the System Preferences rebuild, since it's worked fine for 20+ years of OS X before that.
However, I've noticed, there's not much point in changing it. Showing the desktop is a waste of screen real estate because of the generations of abuse of desktop shortcuts. Even if you are careful, it becomes a cluttered wasteland (on Windows anyways). I just learned to never use the desktop for anything and always have windows up on various monitors.
https://discussions.apple.com/thread/256028948?answerId=2613...
I also ran into it with 15.4.0 and worked around it by creating an image that's the solid color I like to use. It turns out that the system-supplied solid color options are themselves just 128x128 pixel PNG images too:
https://discussions.apple.com/thread/256028948?answerId=2613...
Deleted Comment
Just offer a background picker, and have it generate a 1x1 PNG, for the color selected. Just like that, you can use the image background code path for everything; and the code for generating the image almost certainly exists in the OS already. Maybe add some metadata to the generated image to filter it from the images picker, and you're done.
Quite a capable machine for my uses.
Not supported in Windows 11. Maybe with some additional config? Can’t be bothered with something hat might turn out to be fragile and need more maintenance than I can be bothered with. That’s a young man’s gane.
Ok, I’m about due to give Linux another tickle anyways.
Hmm, which distro… can always give a few a spin.
Keep it simple, Pop!_OS.
Installed fast, no issues, runs fine, stable. Seems entirely usable.
Customisations? Nah, keep it simple.
I’ll set a black background though.
Nope.
Switching to upstream (Ubuntu) with KDE would probably be more your speed.
win11 ltsc works perfectly on it. With a solid background :D
Dead Comment
Dead Comment
Teams not loading due to security issues, but notifications coming through with full content of messages. Ability to type a handful of words in cloud version of Word (or paste full documents) before security check catches up and requires me to set up a sensitivity label. Etc.
It mostly tells me about MS doing very bad software architecture for web apps, though apparently the desktop apps are not immune to it either.
However when you're inside a note (which BTW, can also be converted into checkboxes, aka very simple TODOs), Google Keep, the note taking app from search giant Google, doesn't have search functionality for that specific note.
Besides the many small bugs, sometimes the missing functionality in Google apps is mind boggling.
[1] https://photos.google.com/
Makes me think it must have something to do with their corporate culture and how they work, since their developers, to my knowledge, never had a reputation for being weak. Maybe it's just because they have such a gigantic user base for everything they do, that 80% is the best business value. Though I doubt that a bit, as a third party developer, it makes me avoid their products where I can.
I was thinking about something similar recently. 80% of features take 20% of the time. For the hobby stuff I program, I want to make the best use of my time, so I skip the last 20% of features to make more room for stuff that matters.
I call it PDD: Pareto-Driven Development. Looks like you think Microsoft is similar.
Like in this article alone, that's a change that should never be made. It's an understandable bug but it's indicative of carelessness, both managerial (how does this hold up to code review?) and QA (why is no one asking why login splashes take 30 seconds?).
Only when they "release" it. And then they start again from 10%. /s
I'm certain that some idiotic change just like the ones suggested in the article destroyed this perfectly working feature, and nobody is bothered to fix it because it would impact the latest harebrained scheme to make my 10 year laptop do AI in its sleep and suggest helpful ads about the things it couldn't help overhear while "sleeping".
Most of my computers and friends computers have been ASUS though, maybe that is a connection.
(Windows user since 3.11 but I don't think those had sleep modes :-)
Now you have to guess whether the software has really loaded or not before you start using it.
I could understand it if your device needed special access (VPN to prod etc), but you usually can't do that either from the dev machines - and need to first connect to a virtual machine (via browser or rdp) to be able to do that...
and it has migrated to web apps today - where doing something causes the UI to show a loading/progress wheel, but it takes forever in actuality (or on start up of the webpage, you get a blank screen with placeholder bars/blurred color images etc).
And this is the so-called responsive design...
I’m not sure if this was meant to be a pun, but “responsive design” has nothing to do with how quickly a UI loads. It’s about adapting to different screen sizes.
I've learned to use default configurations pretty much everywhere. It's far too much of a hassle to maintain customizations, so it's easiest to just not care. The exception is my ~50 lines of VS Code settings I have sync'd in a mysterious file somewhere that I've never seen, presumably on github's servers, but not anywhere I can see?
Is it? The vast majority of the time, I change settings/set things up the way I want, and then... leave them for literally years. Hell, I can directly restore a backup I have of Sublime Text from years ago and my customizations will work.
Somewhere along the way I lost interest in customizing the OS. These days I routinely switch between MacOS, Windows and various Linux flavors on lots of computers. The only thing I may customize is I write my .vimrc from memory.
On my Android phones, I change the wallpaper and I disable animations. Otherwise, stock everything.
Now that I think about it, it can't be the time saved, surely I waste more time on HN. It likely correlates more with using computers for work as opposed to for fun and learning. Even the learning I do these days is rather stressful - if I can steal an hour or two on the weekend, I feel lucky, so spending time to customize the environment seems like a waste.
Maybe if life slows down, I'll find joy in customizing my OSes again.
The hard part of maintaining a config is that there's no such thing as cost-free usage, it always takes a mental toll to change a config, to learn a new config, to remember which configs are in use and what they do, to backup configs, or at least to setup and maintain a config-auto-backup flow.
By far, the easiest mental model is just learning how everything works out of the box, and getting used to it. Then again, sometimes what people want is to do things the hard way for its own sake. That's probably part of why I kept going back to C over and over for so many years.
Just your regular reminder that nix is good actually.
"I have a bug, you can get a full VM that reproduces it with 'nixos-rebuild build-vm --flake "github:user/repo#test-vm" && ./result/bin/run-*-vm'"
And the code producing that VM isn't just a binary blob that's a security nightmare, it's plain nix expressions anyone can read (basically json with functions).
And of course applying it to a new machine is a single command too.
(Would it be pedantic of me to say that I receive my fair share of bug reports on nix code I maintain, and when someone sends me their entire nixosConfig the very first thing I do is punt it back with a "can you please create a minimal reproducible configuration"? :D but your point stands. I think. I like to think.)
Deleted Comment
He surely didn't use any Microsoft product. /s
Would have been easier to stick with the pixel density we had.
Oh, and we have to wait a frame to see everything because of compositing that I still don't quite understand what it's supposed to do? Something something backing store?
Yes, xrandr -scale. Works fine for everything. Even better than Windows (which, for some reason, only scales some programs, not all)
More recently, long after I stopped using Windows but still many years ago, I was reading an article about Arthur Whitney. It had a photo which seemd to be at home, maybe in a furnished garage, and in the background was a desktop computer running Windows. The only window open was a cmd.exe. I am not suggesting anything. It is just something I always remember.
Perusing some recent Microsoft documentation I noticed this:
https://learn.microsoft.com/en-us/windows/configuration/shel...
Because they made it a runtime thing - "components just have to remember to do this", the code structure itself affords this bug.
There was a similar bug at facebook years ago where the user's notification count would say you had notifications - and you click it, and there aren't any notifications. The count was updated by a different code path than the code which inserted notifications in the list, and they got out of sync. They changed the code so both the notification count & list were managed by the same part of the system, and the all instances of the bug went away forever.