• 1 Post
  • 18 Comments
Joined 9 months ago
cake
Cake day: September 27th, 2023

help-circle





  • Ah, somehow I didn’t see 18 there and only looked at 17. Thanks!

    I tried pulling just the one package from the sid repo, but that created a cascade of dependencies, including all of llvm. I was able to get those files installed but not able to get clinfo to succeed. I also tried installing llvm-19 from the repo at https://apt.llvm.org/, with similar results. clinfo didn’t throw the fatal errors anymore, but it didn’t work, either. It still reported Number of devices 0 and OpenCL-based tools crashed anyway. Not with the same error, but with something generic about not finding a device or possibly having corrupt drivers.

    Should I bite the bullet and do a full ugprade to sid, or is there some way to this more precisely that won’t muck up Bookworm?





  • Oh yes, definitely. I think this is why Mozilla has not made this the default behavior in Firefox; there will always be the risk of false-positives breaking copied links, so it’s important that people know that there’s some kind of mutation happening.

    ClearURLs uses a JSON file with site-specific regex patterns and rules. In theory I could customize this for myself, or better yet submit a pull request on their GitHub. If I have time I’ll look into it.


  • Personally, I have found this feature to be too limited. I still use the ClearURLs extension, which is more effective in my experience.

    However, neither one is a silver bullet. Here’s an example I just took from Amazon (I blocked out some values with X’s):

    Original URL:
    https://www.amazon.com/Hydro-Flask-Around-Tumbler-Trillium/dp/B0C353845H/ref=XXXX?qid=XXXXXXXXXX&refinements=p_XXXXXXXXXXXXX&rps=1&s=sporting-goods&sr=XXX

    Using Firefox’s “copy link without site tracking” feature:
    https://www.amazon.com/Hydro-Flask-Around-Tumbler-Trillium/dp/B0C353845H/ref=XXXX?qid=XXXXXXXXXX&refinements=p_XXXXXXXXXXXXX&rps=1&s=sporting-goods

    Using ClearURLs:
    https://www.amazon.com/Hydro-Flask-Around-Tumbler-Trillium/dp/B0C353845H?refinements=p_XXXXXXXXXXXXX&rps=1

    The ideal, canonical URL, which no tools I’m familiar with will reliably generate:
    https://www.amazon.com/dp/B0C353845H

    Longer but still fully de-personalized URL:
    https://www.amazon.com/Hydro-Flask-Around-Tumbler-Trillium/dp/B0C353845H

    If anybody knows a better solution that works with a wide variety of sites, please share!


  • This is correct, albeit not universal.

    KDE has a predefined schedule for “release candidates”, which includes RC2 later this month. So “RC1” is clearly not going to be the final version. See: https://community.kde.org/Schedules/February_2024_MegaRelease

    This is at least somewhat common. In fact, it’s the same way the Linux kernel development cycle works. They have 7 release candidates, released on a weekly basis between the beta period and final release. See: https://www.kernel.org/category/releases.html

    In the world of proprietary corporate software, I more often see release candidates presented as potentially final; i.e. literal candidates for release. The idea of scheduling multiple RCs in advance doesn’t make sense in that context, since each one is intended to be the last (with fingers crossed).

    It’s kind of splitting hairs, honestly, and I suspect this distinction has more to do with the transparency of open-source projects than anything else. Apple, for example, may indeed have a schedule for multiple macOS RCs right from the start and simply choose not to share that information. They present every “release candidate” as being potentially the final version (and indeed, the final version will be the same build as the final RC), but in practice there’s always more than one. Also, Apple is hardly an ideal example to follow, since they’ve apparently never even heard of semantic version numbering. Major compatibility-breaking changes are often introduced in minor point releases. It’s infuriating. But I digress.



  • hersh@literature.cafetoLinux@lemmy.mlIs anyone using awk?
    link
    fedilink
    arrow-up
    24
    arrow-down
    1
    ·
    6 months ago

    All the time. Not always by choice!

    A lot of my work involves writing scripts for systems I do not control, using as light a touch as is realistically possible. I know for a fact Python is NOT installed on many of my targets, and it doesn’t make sense to push out a whole Python environment of my own for something as trivial as string manipulation.

    awk is super powerful, but IMHO not powerful enough to justify its complexity, relative to other languages. If you have the freedom to use Python, then I suggest using that for anything advanced. Python skills will serve you better in a wider variety of use cases.




  • I used to run Tumbleweed with KDE on my Nvidia system. I found the rolling release structure of Tumbleweed to cause extra work for me, because kernel updates came frequently and occasionally broke the Nvidia drivers. As a workaround, I ended up pinning my kernel to an old version.

    Nvidia drivers have been at least a little troublesome on every distro I’ve used, particularly with the additional CUDA libraries.

    One nice thing about Suse is that it uses BTRFS by default, and you can use snapper to revert your whole system if something goes wrong. So if Nvidia shits the the bed after an update, it’s easy to roll back. Most distros default to ext4 and do not have snapshot support by default, which feels like living in the stone age to me after using Suse and BTRFS.

    Of course you CAN set up BTRFS and snapshots in any distro, but that’s a lot to ask for a beginner with Linux. I strongly recommend choosing a distro that does that for you, like Suse.



  • I feel this.

    Back in the 90s, there was a fantastic paint program for Mac called ColorIt! (The exclamation point is part of the name, though this is the last time I will respect that because it’s obnoxious; lookin’ at you, Yahoo!*)

    It was a commercial product, but ColorIt 2.3 was eventually released as freeware after newer major versions were released for sale. 2.3 was everything I needed, and while I did try ColorIt 4.0, it didn’t click with me the way 2.3 did. At the time I felt like they bowed to the pressure of Adobe’s success and instead of playing to their unique strengths, they made ColorIt’s UI a bit too much like Photoshop. So I stuck with version 2.3.

    By the time Mac OS X came around, ColorIt was no longer in active development. But OS X had the “Classic” environment, something akin to an OS 9 VM tightly integrated into OS X. Classic apps didn’t look or feel like native OS X apps, and running Classic came with a heavy RAM burden. But I did it anyway, because ColorIt 2.3 was da bomb.

    I continued using ColorIt 2.3 up until Apple killed support for Classic in 10.6 Snow Leopard.

    At that point, the intrepid developers came out of hiding and created a Carbon port of ColorIt 4.5 that could run natively on OS X. It was Carbon-only, which meant that it it didn’t run natively on Intel Macs, but it did run thanks to Apple’s Rosetta compatibility layer — at least until Apple axed that as well.

    If I ever get into pixel art again, I’ll probably run ColorIt 2.3 again in an OS 9 VM with Sheepshaver or whatever works best nowadays.

    *That exclamation point is strictly to emphasize my disdain for Yahoo.