- 337 Posts
- 22 Comments
ylai@lemmy.mlOPto Linux@lemmy.ml•Schenker shows off a Linux laptop prototype with Snapdragon X Elite at Computex 202420·1 year agoLikely due to being a prototype. Production laptops from Tuxedo tend to have the “TUX” penguin in a circle logo on the Super key by default. They also have been offering custom engraved keyboard (even with the entire keyboard engraved from scratch to the customer’s specifications) as added service, so probably there will be suppliers or production facility to change the Super key.
By the way, there was one YouTube channel that ended up ordering a laptop with Windings engraving from them: https://youtu.be/nidnvlt6lzw?t=186
ylai@lemmy.mlOPto Open Source@lemmy.ml•Open Source Initiative tries to define Open Source AI1·1 year agoHow does this analogy work at all? LoRA is chosen by the modifier to be low ranked to accommodate some desktop/workstation memory constraint, not because the other weights are “very hard” to modify if you happens to have the necessary compute and I/O. The development in LoRA is also largely directed by storage reduction (hence not too many layers modified) and preservation of the generalizability (since training generalizable models is hard). The Kronecker product versions, in particular, has been first developed in the context of federated learning, and not for desktop/workstation fine-tuning (also LoRA is fully capable of modifying all weights, it is rather a technique to do it in a correlated fashion to reduce the size of the gradient update). And much development of LoRA happened in the context of otherwise fully open datasets (e.g. LAION), that are just not manageable in desktop/workstation settings.
This narrow perspective of “source” is taking away the actual usefulness of compute/training here. Datasets from e.g. LAION to Common Crawl have been available for some time, along with training code (sometimes independently reproduced) for the Imagen diffusion model or GPT. It is only when e.g. GPT-J came along that somebody invested into the compute (including how to scale it to their specific cluster) that the result became useful.
ylai@lemmy.mlOPto Open Source@lemmy.ml•Open Source Initiative tries to define Open Source AI2·1 year agoThis is a very shallow analogy. Fine-tuning is rather the standard technical approach to reduce compute, even if you have access to the code and all training data. Hence there has always been a rich and established ecosystem for fine-tuning, regardless of “source.” Patching closed-source binaries is not the standard approach, since compilation is far less computational intensive than today’s large scale training.
Java byte codes are a far fetched example. JVM does assume a specific architecture that is particular to the CPU-dominant world when it was developed, and Java byte codes cannot be trivially executed (efficiently) on a GPU or FPGA, for instance.
And by the way, the issue of weight portability is far more relevant than the forced comparison to (simple) code can accomplish. Usually today’s large scale training code is very unique to a particular cluster (or TPU, WSE), as opposed to the resulting weight. Even if you got hold of somebody’s training code, you often have to reinvent the wheel to scale it to your own particular compute hardware, interconnect, I/O pipeline, etc… This is not commodity open source on your home PC or workstation.
ylai@lemmy.mlOPto Open Source@lemmy.ml•Open Source Initiative tries to define Open Source AI2·1 year agoThe situation is somewhat different and nuanced. With weights there are tools for fine-tuning, LoRA/LoHa, PEFT, etc., which presents a different situation as with binaries for programs. You can see that despite e.g. LLaMA being “compiled”, others can significantly use it to make models that surpass the previous iteration (see e.g. recently WizardLM 2 in relation to LLaMA 2). Weights are also to a much larger degree architecturally independent than binaries (you can usually cross train/inference on GPU, Google TPU, Cerebras WSE, etc. with the same weights).
There has been:
ylai@lemmy.mlto Memes@lemmy.ml•I'm giving them a year until lifetime licenses start to mean nothing.61·1 year agoGIMP is a special case. GIMP is being getting outdeveloped by Krita these days. E.g.:
https://gitlab.gnome.org/GNOME/gimp/-/issues/9284
Or compare with:
https://www.phoronix.com/news/Krita-2024-GPUs-AI
GIMP had its share of self inflicted wounds starting with a toxic mailing list that drove away people from professional VFX and surrounding FilmGimp/CinePaint. When the GIMP people subsequently took over the GEGL development from Rhythm & Hues, it took literally 15 years until it barely worked.
Now we are past the era of simple GPU processing into diffusion models/“generative AI” and GIMP is barely keeping up with simple GPU processing (like resizing, see above).
From my own statistics how many I feel worthy posting/linking on Lemmy, the most direct alternative to Kotaku is Eurogamer. PCGamer, PCGamesN and Rock Paper Shotgun are occasionally OK, but you have to cut through a lot of spam and clickbait (i.e. exactly this “50 guides per week” type of corporate guidance). Not sure if this is also the state that Kotaku will end up in. The Verge sometimes also have good articles, but the flood of gadget consumerism articles there is obnoxious.
In the beginning, only privileged ones will be allowed to run in pass-through mode. But goal/roadmap calls for all FUSE filesystems eventually to have this near-native performance.
Well, if you have a constructive suggestion which site to link instead regarding kernel developments, I am all ears:
- Not sure that raw commits are readable or have sufficient context for non kernel development readers here
- LWN, particularly timely/kernel development news there, has gone mostly paywall, and there will be (legitimate) complaint if I link articles needing a LWN subscription
Not sure what called for this blatant personal attack. My post history speaks for itself, quite in comparison to yours. And Phoronix is well-known Linux website, and its test suite is in fact even referenced in various regression tests/patches in LKML (also not sure what/if any kind of kernel development you have done).
ylai@lemmy.mlOPto Gaming@lemmy.ml•Geoff Keighley Finally Tweets About Layoffs, And The Timing Is Awfully Convenient101·1 year agoThere is pre-existing context and criticism. And it is not about, or just being the perception of “this journalist”:
https://www.theverge.com/23992402/geoff-keighley-the-game-awards-layoffs
https://videogames.si.com/features/games-industry-deserves-better-than-geoff-keighley
https://www.inverse.com/gaming/the-game-awards-2023-needs-to-acknowledge-industrys-lay-offs-problem
https://dotesports.com/the-game-awards/news/the-game-awards-layoffs-developers-no-respect
The problems also goes beyond just the layoffs, but his overt coziness and preferential treatment of large studios, over even the ones that actually won the award he is presiding over, and are supposed to be celebrated:
https://insider-gaming.com/geoff-keighley-shows-cowardice-at-the-game-awards/
https://www.eurogamer.net/the-game-awards-speeches-were-too-short-geoff-keighley-admits
ylai@lemmy.mlOPto Gaming@lemmy.ml•Investigative Report: The Day Before – Fraud or Megalomania? (German audio with English subtitle)2·1 year agoThere are now summaries from non pay-walled (and English) press: https://www.eurogamer.net/new-the-day-before-report-alleges-employees-fined-for-making-mistakes
ylai@lemmy.mlto Linux@lemmy.ml•Docker team is considering distributing Docker Desktop as a Flatpak and Snap11·1 year agoYes. If you mean “CLI” as for e.g. pacman install, it is a GUI (Electron) application, so I expect will install straight from e.g. KDE Discover and then run without you touching the shell.
ylai@lemmy.mlto Linux@lemmy.ml•Docker team is considering distributing Docker Desktop as a Flatpak and Snap21·1 year agoInstalling podman-compose with the immutable filesystem is fairly straight forward, since it is just a single Python file (https://github.com/containers/podman-compose/blob/devel/podman_compose.py), which you can basically install anywhere in your path. You can also first bootstrap pip (
python3 get-pip.py --user
withget-pip.py
from https://github.com/pypa/get-pip) and then dopip3 install --user podman-compose
.
ylai@lemmy.mlto Linux@lemmy.ml•Docker team is considering distributing Docker Desktop as a Flatpak and Snap151·1 year agoThere might be several misunderstandings:
- Docker Desktop ≠ Docker Engine, and I think what you (and several in this thread) are thinking is actually Docker Engine. Docker Desktop ultimately includes a Docker Engine inside, but it does not appear you need that virtual machine (e.g. running non-Linux code). See: https://docs.docker.com/desktop/faqs/linuxfaqs/#what-is-the-difference-between-docker-desktop-for-linux-and-docker-engine
- Docker Desktop is based on KVM, which already works with Flatpak. So this is not something new. For example, GNOME Boxes is available as Flatpak and provides a way to run KVM guests in SteamOS.
- Starting with version 3.5 (the current stable) SteamOS already includes Podman with the default installation. And running the daemon-y Docker Engine “bare metal” is not going to be any easier with the immutable filesystem. While Docker Desktop solves this by using KVM, it adds another layer with performance loss, vs. just running Podman containers.
So what you want is already available, and no Docker Desktop is actually needed.
AMD’s support for AI is just fine
This is quite untrue, especially if you do actual research and not just run other people’s models. For example, ROCm is missing in many sparse autograd frameworks, e.g. pytorch_sparse, or having a viable alternative to Nvidias MinkowskiEngine. This is needed if you do any state-of-the-art convnets with attention-like sparsity.
ylai@lemmy.mlOPto Selfhosted@lemmy.world•Intel Begins Sorting Out SR-IOV Support For The Xe Kernel Graphics DriverEnglish31·2 years agoYes. But one should also note that only a limited range of Intel GPU support SR-IOV.
ylai@lemmy.mlOPto Open Source@lemmy.ml•Joint Statement by Free Software Foundation Europe and Software Freedom Conservancy Regarding Eben Moglen and Software Freedom Law Center111·2 years agoFSFE’s statement:
Some related personal blogs I noticed:
Just for reference, a few years back, (ex-Microsoft) David Plummer had this historical dive into the (MIPS) origin of the blue color, and how Windows is not blue anymore: https://youtu.be/KgqJJECQQH0?t=780