

IBM/Red Hat maybe since it’s a US company. HOWEVER we went through this with PGP already and the infamous RSA Dolphin.
So could they try? Yeah. Would it work? I don’t know.
Just a dad with a sysadmin hobby … leaving reddit
IBM/Red Hat maybe since it’s a US company. HOWEVER we went through this with PGP already and the infamous RSA Dolphin.
So could they try? Yeah. Would it work? I don’t know.
L2ARC only does metadata out of the box. You have to tell it to do data & metadata. Plus for everything in L2ARC there has to be a memory page for it. So for that reason it’s better to max out your system memory before doing L2ARC.
It’s also not a cache in the way that LVMCACHE and BCACHE are.
At least that’s my understanding from having used it on storage servers and reading the documentation.
I used to do this all the time! So in terms of speed bcache is the fastest, but it’s not as well supported as lvm cache. IMHO lvm cache is plenty fast enough for most uses.
Is it going to be as fast as a NVME ssd? Nope. But it should be about as fast as a SATA ssd if not a little slower depending on how it’s getting the data. If you’re willing to take that trade off it’s worth it. Though anything already cached is going to be accessed at NVME speeds.
So it’s totally worth it if you need bigger storage but can’t afford the SSD. I would go bigger in your HDD though, if you can. Because unless you’re accessing more than the capacity of your SSD frequently; the caching will work extremely well for both reads and writes. So your steam games will feel like they’re on a SSD, most of the time, and everything else you do will “feel” snappy too.
OP said they don’t want/can’t use the built in paid sync service and wants to self host it
I’ve been putting everything behind Tailscale. I don’t see any reason to make it public unless you’re planning on sharing it with the public.
Why? And what would be a replacement for it?
Restic, it has native S3 compatibility and when you combine with something like B2 it makes amazing offsite storage so you can enjoy the tried and true 3-2-1 backup strategy.
Also fedora magazine did a few posts on setting it up with systemd that makes it SUPER EASY to get going if you need a guide.
I have an ansible role that configures it on everyone’s laptops so that they have local, NAS, and remote, B2, backup locations.
Works like a charm for the past 8+ years.
MacOS, nearly everyone who does anything with development or ops is using a MacBook. Though lately more “normal” employees have been getting MacBooks too.
Waaaaay better.
Restic allows you to make dedupe snapshots of your data. Everything is there and it’s damn hard to loose anything. I use backblaze b2 as my long term end point / offsite… some will use AWS glacier. But you don’t have to use any cloud services. You can just have a restic repository on some external drives. That’s what I use for my second copy of things. I also will do an annual backup to a hard disk that I leave with a friend for a second offsite copy.
I’ve been backing up all of my stuff like this for years now. I used to use BORG which is another great tool. But restic is more flexible with allowing multiple systems to use a single repository and has native support for things like B2 that BORG doesn’t.
We also use restic to backup control nodes for some of supercomputing clusters I manage. It’s that rock solid imho.
You could also just grab a wire guard configuration and use it too. They provide them along side the openvpn configs
To be honest, there’s a few good comments linking to scripts and methods here to batch convert them on a windows pc/vm. That’s the best way to go.
To add on to their comments. If you’re just interested in preserving them then maybe printing them to pdf, specifically pdf/a, would be my approach once you got them opened.
I’ll leave this one here for someone:
You can tunnel L2 over OpenVPN. Just bridge your interfaces in both sides and it works.
That way if you need to provision a VOIP phone or just have something NetBoot remotely. Not that I recommend doing that…
TBH have you tried just basic git? There’s a web interface built into git itself and you can use ssh for your repositories. It’s simple and just works. If you need a faster web interface there’s also cgit. There’s no bells and whistles either. Just configure ssh, drop your repos in /srv and get to work.
If you need more that just standard basic git the. The other suggestions here are great especially forgjo!
Once I figured out how to netboot the os into memory that’s how I run all my nodes :)
I use backblaze b2 for my storage. I use restic to backup everything to it. It works well and I’ve had it going for YEARS at this point. For things I could never replaced, like photos, I use external drives in addition to B2. Everyone knows that if something happens and we need to leave to just grab the drive that is stuck to the wall and the family photos will be safe.
My though process goes like this, everything backups to my home server. I have snap shots of the data on a normal basis. So if I need to get something back, going to a snap shot is pretty simple. If for some reason my server(s) just stopped existing for some reason I could pull it back from B2. I’ve only had to actually restore from B2 a handful of times and it was worth it.
Yeap! That’s most of the reasons why, IMO, we didn’t see the adoption of SUSE like we did with RedHat based stuff.
I’ve been a huge supporter of RH over the years. After IBM bought them I was cautious but believed that RH would continue mostly the same. Then they killed CentOS. I understood the reasons and the arguments that Stream made sense for them. I was highly annoyed and disappointed with that decision but Rocky showed up and it was okay.
This final round of fuckery really shows me that the RH we all supported is gone. They burned down nearly all the good will they built up. Locking the source behind a paywall was just the final insult as far as I’m concerned.
I don’t know how much this will hit their bottom line but I suspect it’s going to have some kind of financial impact just from the number of people like me who used CentOS/Rocky as a gateway to RHEL in prod. I suspect they know and just don’t care.
With that being said I’m looking at how they react to over the next year or so. If they’re going to be more of an asshat then I’ll start retooling for another distribution. I here nix is pretty good these days…
I’ve abused syncthing in some many ways migrating servers and giant data sets. It’s freaking amazing. Though it’s been a few years since I’ve used it. Can only guess how much better it’s gotten.