moonpiedumplings

joined 1 year ago
[–] moonpiedumplings@programming.dev 19 points 1 week ago* (last edited 4 days ago) (1 children)

Ubuntu in WSL comes with systemd enabled. Debian doesn't, and you have to enable it yourself.

That's why I chose to have people use Ubuntu in WSL, despite the other downsides. One less step to setup a Linux environment on Windows makes the process smoother.

Wish I could transcend into declarativity but the thread’s nix survivor ratio is grim

Yeah lol.

I will say, that for my server, I decided to use kubernetes + fluxcd for declaratively. My entire kubernetes "state" is declared in a git repo, and this is the popular, industry standard for things like this, called GitOps. It makes it very easy to add an app, since it's just adding a folder + some new config files. And unlike Nix, Kubernetes and Flux are very well documented with much tooling as well. Nix doesn't really have a working LSP or good code autocomplete, but with kubernetes, I can just start typing in a yaml file and then hit tab and it spits out the template for me. Code autocompletion with kubernetes feels much more similar to the tooling of other, more mature tooling

It's not as declarative as nix though. There are things missing, like OCI containers could theoretically shift if you don't rely on hashes and some other nitpicks. But declarativity is a spectrum, and I feel like, outside of scientific scenarios (think simulations where versioning, hardware, runtime etc being the same is very important), I think many non-nixos solutions are declarative enough.

[–] moonpiedumplings@programming.dev 3 points 1 week ago* (last edited 4 days ago)

Advice online seemed like i needed to basically create a nix flake for the app. I still havent gotten it installed because i have no idea what nix flakes are.

So, the problem is that flakes are technically an "experimental" feature, and thus are not allowed to be included as a primary solution in the official documentation. But, basically everybody uses flakes, so it leads to this crazy documentation split, and is a big part of why documentation on Nix is so part.

Some stuff can only be done with flakes, some stuff only with non-flakes and you have to figure out which is which on your own, while also dealing with the poor documentation for either.

The advice you received was wrong. You could also use a combination of a default.nix file and a shell.nix file to create a package and development environment for your app. But, the documentation is so poor that it's unlikely you will learn this, and figuring out how to do this on your own, is again, a massive time sink.

[–] moonpiedumplings@programming.dev 3 points 1 week ago* (last edited 3 days ago) (2 children)

So, I use Arch, but I don't use the AUR at all. Instead, I use nixpkgs to get stuff (admittedly only like 3 packages) not in the Arch repos.

The main reason for this is the quality of AUR packages. Although I don't really fear a malicious package, I do remember hearing about a package that moved a users /bin to /opt during the install phase.

Something like that is literally impossible with Nix, due to the way that applications aren't really installed to the system. But, nixpkgs also requires some level of vetting the package quality, which is also nice.

I also use nix for managing all my development environments. For example, my blog github repo, has a few nix files at it's root, and you should just be able to type nix-shell in folder, and then you will get an identical environment to me.

declarative rollbackable immutability sounds really freakin’ AWESOME

I have BTRFS snapshots set up, and with grub-btrfs, I can even boot from them and revert to an older kernel (my /boot is stored on BTRFS).

However, I have given up on NixOS, for many reasons. The documentation is very poor, and it's more complexity than it's worth, to make my whole OS reproducible, rather than just my development environments. In addition to that, their are also issues with running certain apps that expect to see a normal FIlesystem Hierarchy, which nix does not provide. Although you can work around this with stuff like steam-run or creating a fake FHS using nix, I would rather not play that game.

But, considering I installed some stuff in an Ubuntu 22 distrobox recently, because that was what VScode and Unity official provide repos for, maybe this doesn't really matter. You can probably use distrobox on Nixos, but I've seen issues about GPU acceleration with distrobox (and other non-nix apps) as well.

EDIT: I lied, I use the chaotic aur for some things.

[–] moonpiedumplings@programming.dev 4 points 1 week ago* (last edited 4 days ago)

OP seems to be trying to install older projects, rather than creating a new project.

Oops... my bad. In my earlier comment I assumed that this would be a Fedora/Ublue based distro, rather than an Arch one. Arch doesn't have RPM ostree either (which makes me dislike it as a choice for an immutable distro).

But, it's highly likely that with the steam deck and other projects, there is already an ecosystem for immutable Arch, and a minimal base system to start is advantageous, as Possibly Linux said.

[–] moonpiedumplings@programming.dev 12 points 1 week ago (1 children)

Yes. Firstly, it's about release cycles. Centos Stream is a rolling release distro (although it rolls very, very slowly). But what this means, is that there isn't a true guarantee of application/ABI/API compatibility between current versions of Centos Stream and future versions.

In constrast to this, Centos 8 and previous were complete clones of Red Hat Enterprise Linux, which was a stable release distro. During the 10 year lifecycle of each RHEL release, there was a guarantee certain application/ABI/API compatibility not changing, which is what stability in the Linux/software world is defined as.

Centos 8 was a free alternative, for institutions unwilling, or unable to pay for RHEL stable releases. But, with the death of Centos, an alternative was needed. Alma Linux, Rocky Linux, and Scientific Linux (designed for labs and universities), were rebuilds of RHEL. This meant that, they would take RHEL's open source code, and recompile it and distribute it in a way that guaranteed application/ABI/API compatibility with RHEL, for the same lifecycle of a RHEL release.

So Alma Linux and Rocky Linux fill that gap... but recently, RHEL said that they are adjusting policies to make it much harder for people to make rebuilds (likely targeting Oracle Linux, which is a RHEL rebuild), but this change may affect Alma and Rocky as well.

Rocky said they were going to keep bug-for-bug compatibility, like they used to, but Alma says they are going to do something different. Although they still intend to be ABI compatible, Alma has decided to make some changes to the base system, such as reimplimenting and continuing to support things that Red Hat saw unfit to continue existing in RHEL. One example of this is SPICE, which is a graphics protocol used for low latency display of virtual machines. It had many usecases, and I am very excited to see it back in a distro in the Red Hat ecosystem.

[–] moonpiedumplings@programming.dev 32 points 1 week ago (2 children)

https://help.kagi.com/orion/faq/faq.html#oss

We're working on it! We've started with some of our components and intend to open more in the future.

The idea that "open-source = trustworthy" only goes so far. For example, the same tech company that offers a popular open-source browser also has the largest ad/tracking network in history, with that browser playing a significant role in it. Another company with a closed-source browser (using WebKit like Orion) is on the forefront of privacy awareness and technologies in its products.

So, does anyone here remember when all chromium browsers had a secret api that sent extra data to google? Brave, Opera, and Edge got hit by this one, but I think Vivaldi dodged it. They all removed this after they found out, but still...

When it comes to things like browsers, due to the sheer complexity and difficulty to truly audit chromium, I don't really consider chromium to be "open source" in the same sense as many other apps. Legally, you can see and edit the code. But in practice, it's impossible to audit all of it, and the development is controlled by a single corporation who puts secrets in it, or removes features that harm their interests (manifest v3). Personally, I consider Minecraft Java to be closer to open source than chromium is.

To say that:

The idea that "open-source = trustworthy" only goes so far

is really just a cop-out and excuse for not being transparent with their code and what they are doing.

[–] moonpiedumplings@programming.dev 2 points 1 week ago (2 children)

Opensuse doesn't have rpm-ostree. Their immutable offerings are just snapper/btrfs snapshots before changes to the system.

Such a setup is nowhere near as powerful. rpm-ostree can rebase itself based off of a container/oci image. It can layer images on top of eachother. Rather than just tracking when changes happened, it can also track what change happened, in a git style setup.

[–] moonpiedumplings@programming.dev 8 points 1 week ago (1 children)

The whole point of it is that in a truly random system all known patterns should eventually emerge somewhere within it.

So pi (probably) has this property. There are some joke compression programs around this (they don't really work because it takes up more space to store where something in pi is, than storing the thing itself). But it is funny, to think that pi could theoretically hold every past, present, and future piece of information within those digits after the decimal.

https://github.com/philipl/pifs

https://ntietz.com/blog/why-we-cant-compress-messages-with-pi/

No need for AI for that, humans can do it better:

https://youtube.com/watch?v=l7ZUZerGwK4

https://youtube.com/watch?v=zn_rx8Zyl54

If you know where to look, someone already did it.

 

I'm having trouble with networking on linux. I am renting a vps with only one NIC, one ipv4 address, and a /64 range of ipv6 ones. I want to deploy openstack neutron to this vps, but openstack neutron is designed to be ran on machines with two NIC's, one for normal network access, and entirely dedicated to virtualized networking, like in my case, giving an openstack virtual machine a public ipv6 address. I want to create a virtual NIC, which can get it's own public ipv6 addresses, for the vm's, without losing functionality of the main NIC, and I also want the vm's to have ipv4 connectivity. I know this setup is possible, as the openstack docs say so, but they didnt' cover how to do so.

Docs: https://docs.openstack.org/kolla-ansible/latest/reference/networking/neutron.html#example-shared-interface

There is an overview of what you need to do here, but I don't understand how to turn this into a usable setup. In addition to that, it seems you would need to give vm's public ipv4 addresses, in order for them to have internet connectivity. I would need to create a NAT type network that routes through the main working interface, and then put the neutron interface partially behind that, in order for ipv4 connectivity to happen.

I've been searching around for a bit, so I know this exact setup is possible: https://jamielinux.com/docs/libvirt-networking-handbook/multiple-networks.html#example-2 (last updated in 2016, outdated)

But I haven't found an updated guide on how to do it.

view more: ‹ prev next ›