lemmyvore

@lemmyvore@feddit.nl

This profile is from a federated server and may be incomplete. Browse more on the original instance.

lemmyvore,

I was thinking about getting a new phone and then I just picked up another 10 III. It’s honestly just everything I need.

lemmyvore, (edited )

Tasker? I feel like it comes pretty close to what you describe.

lemmyvore,

Wow you got that backwards. They don’t do any of that for the sake of Nouveau or Vulkan or Wayland or whatever. They don’t care what people use their open scraps for.

They open up the minimum they can get away with because it’s ultimately meaningless — their proprietary stuff is still hidden away and it’s not like you can use the parts they open with anything else.

This, btw, applies to AMD and Intel too. The only choice you get with proprietary hardware that you have to use (like GPUs) is whose dick you want to suck. They’re not your friend and they won’t let community pressure then into decisions.

lemmyvore,

I mean I’m fully with you on the fact screen autodetect isn’t stellar on X but there’s no need to exaggerate with “2 or 3 scripts”. It’s one xrandr command.

lemmyvore,

They’ll recant after their usage drops to a fraction. This move makes zero sense no matter how you look at it. As a generalist distro it’s too early to drop X.

If they want to become a niche distro whose only claim to fame is “we only pack Plasma 6”, big whoop, like there’s any shortage of that. What kind of distro defines itself by what it does not offer? And is that the kind of distro that Fedora aims to be?

lemmyvore,

And I’m sure all the other people using 6 monitors on 2 GPUs at the same time will appreciate it.

Seriously, how common is such a scenario that you’d even mention it in this context?

lemmyvore,

A sizable percentage of Linux users own Nvidia cards and “just buy something else” is not realistic, for many reasons.

Wayland will eventually have to support Nvidia one way or another. If they’re seriously considering not doing that I would not bet on its future.

lemmyvore,

Yeah I don’t get why some people would think sticking to X is fanboyism. Nobody likes X, let alone love it. Most people’s relation to X is pragmatic, it’s “it works and does everything I need”.

If anything, fanboyism is telling people they have to use Wayland when it doesn’t yet work for what they need it to do.

Just keep improving the damn thing and people will switch when it’s ready. There’s no convincing needed.

lemmyvore,

Nobody’s getting rid of X support. Not for several years.

lemmyvore,

Not supporting Nvidia cards will make Wayland unusable for at least half the Linux desktop users, probably more. Stats I recall range from 50-75%.

“Just buy non-Nvidia” is not, I repeat, a simple option. Lots of people stick with old GPU models because the price/performance ratio has gone out the window and they cannot afford to drop hundreds or thousands on one. Many others have Nvidia in their laptops.

There’s nothing preventing Wayland from working with Nvidia except the political insistence that it be open sourced. Which Nvidia is not interested in, never was, and never will be. And it’s a red herring to begin with.

TLDR either Wayland bends their stance on open source or their adoption will be severely limited.

Me vs my ISP

So I was looking into getting port forwarding set up and I realized just how closed-off the internet has gotten since the early days. It’s concerning. It used to be you would buy your own router and connect it to the internet, and that router would control port-forwarding and what-have-you....

lemmyvore,

Are you trying to offer a port for peer sharing (XDCC/BT)? I’ve never tried using it like this but I think Tailscale Funnel could work.

It’s a sort of reverse VPN, I guess you could call it. Tailscale maintains the public IP and when someone connects to your advertised port they tunnel it to you through (encrypted) WireGuard. It passes through NAT because connections are outgoing to their servers.

The catch is that wireguard is easily detectable through deep packet inspection so if your ISP is a real asshole they can kill the connections, but if they go that far then NAT traversal is the least of your worries.

lemmyvore,

They can if the license you granted them says they can. Read it. These platforms usually make you grant then extensive rights. Yes they don’t own the content but given such broad permissions it makes very little practical difference.

lemmyvore,

And yet Microsoft made Copilot, and there are currently lots of clueless programmers out there using it to inject code with god knows what licenses into their company’s software.

lemmyvore,

You own the original, which you’ve written on your pc or phone. But the one that ends up on the website is a copy, on which you’ve granted the website owner a non-revokable license to do with as they please ie. a copy-right.

lemmyvore,

They’re not going to be absolved of copying code verbatim without following its license.

lemmyvore,

Not everybody has the knowledge to deal with Linux. A product line TrueNAS or Unraid has a friendly GUI that can be used by a non-technical user.

lemmyvore,

Yeah I expect acting as SSD bays could become popular in the future if SSD prices drop low enough. Although they might be M.2 bays by then.

I have a bunch of old 60 GB 2.5" SSDs around but they’re so small it’s not worth bothering to set up an array of them. Plus they’re more useful individually for stuff like upgrading an old laptop, portable USB storage or installing Windows the one time in three years I need it.

In the meantime I’ve liberated the 2x HDD cage from a Define C Mini’s shroud and mounted it on the floor in a fan slot.

lemmyvore,

if you use the same repository with multiple machines, they have to rebuild their index every time they switch

I’m a beginner with Borg so sorry in advance if I say something incorrect l. I backup the same files to multiple distinct external HDDs and my solution was to use distinct repos for each one. They have different IDs so the caches are different too. The include/exclude list is redundant but I can live with that.

Is this Seagate Exos drive too good to be true?

I found this its the cheapest 10TB Exos drive on Newegg and looking to buy 4 of them. I will be putting them in my NAS that I use for my media library and pc backups. The price I’m posting this is $130, I’m also looking similar Exos drives that are $250 is there a difference? Should I shell up for the more expensive drives?

lemmyvore,

What kind of attributes did you find relevant? I imagine the 19x codes…

I’ve read the Blackblaze statistics and I’m using a tool (Scrutiny) that takes those stats into account for computing failure probability, but at the end of the day the most reliable tell is when a drive gets kicked out of an array (and/or can’t pass the long smart test anymore).

Meanwhile, I have drives with “lesser” attributes sitting on warning values (like command timeout) and ofc I monitor them and have good drives on standby, but they still seem to chug along fine for now.

lemmyvore,

Data centers replace drives when they fail and that’s about it. They don’t care much about SMART data.

lemmyvore,

There is an alternate verification method using an API key to your DNS provider, if it’s a supported one. That method doesn’t need any IP to be assigned (doesn’t care if there are A/AAAA records or where they point because it can verify the domain directly).

deSEC.io is a good example of a good, reputable and free DNS provider that additionally allows you to manage API keys. The catch is that they require you to enable DNSSEC (their mission is similar to Let’s Encrypt, but for DNS).

lemmyvore,

I see that you want to use the cert for intranet apps btw.

What I did was get two LE wildcard certs, one for *.my.dom and one for *.local.my.dom. Both of them can be obtained and renewed with the API approach without any further care to what they actually point at.

Also, by using wildcards, you don’t give away any of your subdomains. LE requests are public so if you get a cert for a specific subdomain everybody will know about it. local.my.dom will be known but since that’s only used on my LAN it doesn’t matter.

Then what I do for externally exposed apps is to point my.dom to an IP (A record) and either make a wildcard CNAME for everything *.my.dom to my.dom, or explicit subdomain CNAME’s as needed, also to my.dom.

This way you only have one record to update for the IP and everything else will pick it up. I prefer the second approach and I use a cryptic subdomain name (ie. don’t use jellyfin.my.dom) so I cut down on brute force guessing.

The IP points at my router, which forwards 443 (or a different port of you prefer) to a reverse proxy that uses the *.my.dom LE cert. If whatever tries to access the port doesn’t provide the correct full domain name they get an error from the proxy.

For the internal stuff I use dnsmasq which has a feature that will override all DNS resolves for anything ending with .local.my.dom to the LAN IP of the reverse proxy. Which uses the *.local.my.dom LE cert for these ones but otherwise works the same.

Is there such a thing as split-screen grep?

I want to run a command and see all of its output on the left hand side, while simultaneously searching/grepping for particular lines on the right hand side. In other words, I want a temporary vertically split screen in my CLI, ideally with scrollback on each side of the split, but where I expect the left hand side to be...

lemmyvore, (edited )

Run rsync, pipe to tee, and redirect the output to a named pipe (mkfifo). Open a second terminal and direct the named pipe into a grep command. Arrange the terminals in whatever way you want.


<span style="color:#323232;">mkfifo mypipe
</span><span style="color:#323232;">rsync | tee mypipe
</span><span style="color:#323232;">grep "denied" < mypipe
</span>

Do you mount an embedded Linux file system to the workstation and use your host scripts or do you SSH/SCP and deal with the limited shell commands?

I’m playing with a couple of routers and comparing proprietary to open source on the same hardware. I miss my .bashrc functions and aliases… and compgen, tree, manpages, detailed help, etc; the little things that get annoying when they are missing....

lemmyvore, (edited )

There isn’t usually much to do on an embedded router other than use its own commands to change settings or manage packages. And if it has enough juice to run more advanced stuff it probably has bash available too.

Anyway, there’s NFS for mounting filesystems remotely. It’s not very complicated, the catch is that the same UIDs and GUIDs on the host must exist on the guest, because it doesn’t do any uid translation. On an embedded system most stuff is owned by root, meaning you’d have to use root on guest too, which may not be a great idea.

Secondly, you can’t run commands over NFS, just manipulate files and I’m not sure that’s something you really need to do a lot of on a router.

Self-hosted VPN that can be accessed via browser extension

Currently I set up Tailscale in my Synology NAS and I can access selfhosted services on my phone using the Android app. I want to use some services in my work PC too but I’m blocked from installing any software. So my question is, is there any solution that allows me to connect to selfhosted VPN via browser extension? (Just...

lemmyvore,

Or you can use the CF Tunnel equivalent from Tailscale, called Funnel.

tailscale.com/blog/reintroducing-serve-funnel

  • All
  • Subscribed
  • Moderated
  • Favorites
  • localhost
  • All magazines
  • Loading…
    Loading the web debug toolbar…
    Attempt #