FOSS in general needs better means of financial support. While the software is free and libre, developer time is not, and ultimately they gotta eat and pay bills. I hope they get positive results and don’t catch much unnecessary flak.
FOSS in general needs better means of financial support. While the software is free and libre, developer time is not, and ultimately they gotta eat and pay bills. I hope they get positive results and don’t catch much unnecessary flak.
My first programming experience, an online class, was in a Linux VM. Linux made programming easy and delightful, Windows always made it a huge pain. As time went on, more of what I did was easier on Linux, and now everything is.
Key detail in the actual memo is that they’re not using just an LLM. “Wallach anticipates proposals that include novel combinations of software analysis, such as static and dynamic analysis, and large language models.”
They also are clearly aware of scope limitations. They explicitly call out some software, like entire kernels or pointer arithmetic heavy code, as being out of scope. They also seem to not anticipate 100% automation.
So with context, they seem open to any solutions to “how can we convert legacy C to Rust.” Obviously LLMs and machine learning are attractive avenues of investigation, current models are demonstrably able to write some valid Rust and transliterate some code. I use them, they work more often than not for simpler tasks.
TL;DR: they want to accelerate converting C to Rust. LLMs and machine learning are some techniques they’re investigating as components.
The moment that shocked me was when printers, network cards, and even motherboard integrated Ethernet didn’t work on Windows without driver downloads but Linux was plug and play. Full reversal of the situation.
If you read carefully this is actually very similar to the Steam news. I doubt Valve or GOG care, but generally the games are “sold” by the publisher as non transferable licenses for you to play them. So the part that matters isn’t up to them.
Note the versions, none of the results give you the official operators page for the current version, 16. They give 9, which went EOL in 2021.
A major caveat I’ve noticed some people misunderstand: it’s corporate CLAs that are problematic. The Apache Foundation also requires contributors sign a CLA, but it’s to provide a legal fail safe and a way to update to say Apache 3.0 if need be one day. Apache’s non profit, open source mission aligns with respecting the rights of contributors and the community. Corporations, on the other hand, not so much.
Codeberg is run off of donations, they have no service contract revenue. Nobody, much less a volunteer, wants to commit to a 5 or 10 year service plan like that, it’s not sustainable for a small project from a non profit.
CLAs can be abusive, but not necessarily. Apache Foundation contributors need to sign CLAs, which essentially codify in contract form the terms of the Apache 2.0 license. It’s a precaution, in case some jurisdiction doesn’t uphold the passive licensing scheme used otherwise. There’s also a relicensing clause, but that’s restricted to keeping in spirit, they can’t close the source.
Have you used it recently? Previous versions I would’ve agreed, but 5.0 was a huge improvement. If I didn’t know, I’d likely have assumed it to be a native feature.
I’ll take a look at Vivaldi’s approach though, I’ve heard good things about those features previously.
If you want vertical tabs with the ability to organize them in trees I suggest the Sideberry extension. It legitimately makes me nervous that the functionality would ever go away, it improves my productivity so much.
You can bookmark trees, collapse them, search them, load/unload them manually, I could go on. It makes it easy to organize dozens or hundreds of tabs. I have some trees for emails, news, forums, projects, etc. When I’m done just fold it up: the top tab bar can hide tabs that aren’t in the active tree you’re using, so you can still navigate the tabs normally.
GNOME always seemed like an odd choice considering how little customization is available. It feels like a prescriptive approach, you will use your computer the way GNOME feels is appropriate, whereas KDE tries to accommodate however you want to use your computer.
3 or 4 years, including on Nvidia machines. I’ll admit it took fiddling to get working awhile ago. Nowadays I use my desktops AMD iGPU as the main display driver and offload the rendering to the Nvidia card for intense programs or games, best of both worlds.
This is almost entirely misdirected. The success of Wikipedia is from its human structures, the technical structure is close to meaningless. To propose a serious alternative you’d have to approach it from a social direction, how are you going to build a moderation incentive structures that forces your ideal outcomes?
Federation isn’t a magic bullet for moderation, alone it creates fractal moderation problems.
Arch for stuff I have physical access to. Nothing’s ever gone wrong, so it’s worth it for the immediate updates and consistency with my other systems. For VPS I use Debian though, occasionally the unstable/Sid branch if I really need the latest updates. There are almost always Debian images available on a VPS.
I think about it like a tree structure for both. With a gui you have to move your mouse around to various places, with a cli each character branches off into another tree. Mathematically you can handle more options faster with a CLI.
Curiosity, back around 2010 before I was a teenager. No clue how I heard about it, but the concept of replacing the entire operating system was fascinating. I figured it must be really good if it was such a well kept secret.
A few years later, when I started to learn programming, Linux was the obvious winner. The online course taught C in a Linux environment, and I was amazed that the default Ubuntu build at the time had everything built in, whereas a Windows equivalent required visual studio and licensing adventures.
It really stuck as a daily driver after Windows 7, where a clear trend emerged: Windows got in my way, Linux got out of my way. Simple as.
Forgejo is my go to, I ran it in a GCP micro instance, which has 768 MB ram and a piddling processor. One of my friends works for a company that had all their devs run a local instance in addition to the main repo, it was that light.
Gitea is the former go to, but gitea was hijacked and stolen from the community by a for profit company. Forgejo is currently a drop in replacement fork, but with added privacy features, future federation options, and a reputable parent organization.
I use netdata, it’s very good at digesting thousands of metrics to sharing actionable. The cloud portion is proprietary, but you can toggle off the data collection. I did turn on the cloud portion though, I get email notifications when something breaks. Might sound counter to the self hosted mantra, but a self hosted monitoring system isn’t very helpful when your own systems go down.
The comments from that article are some of the most vitriolic I’ve ever seen on a technical issue. Goes to prove the maintainer’s point though.
Some are good for a laugh though, like assertions that Rust in the kernel is a Microsoft sabotage op or LLVM is for grifters and thieves.