

The support isn’t exclusivity for native 32 bit cpus, it’s for 32bit libraries that compatibility applications like wine/proton depend on to run 32bit windows executables
The support isn’t exclusivity for native 32 bit cpus, it’s for 32bit libraries that compatibility applications like wine/proton depend on to run 32bit windows executables
99% of the waiting time in my case is either waiting for file copies or waiting on SAP programs to run.
I wish I had low hanging fruit like that to go after.
I copy the install media locally. Although there is probably a noticable performance hot to running my main VM disk over the network.
They’re trying to, but market adoption has said so far that we’re unwilling to tolerate it.
Building a fully functional SAP system just takes that long in raw install time when your process also includes a sufficiently large system copy, and your hardware isn’t bleeding-edge.
It’s a massive application stack
I wrote and maintain a zero-to-working SAP HANA/S4 installer in pure bash.
It takes a redhat compatible from base install to a working, production-ready SAP system in about 5 hours.
It’s like ~9,000 lines of bash
I’m with you that he doesn’t strictly need a gpu, but if the price is right (free from old gaming PC, cheap from a friend’s old gaming PC, cheap old workstation card, etc) I stand by that he probably wants one. A lot less fussy, a lot more capable, nad nvenc does better quality encoding at lower bitrates (and probably less power too if you take into account time spent encoding at full tilt.)
Generally power supplies are the most electrically efficient at 20-60% utilization, so there’s no issue with over-provisioning power, other than the (generally minor) upfront extra cost, which might very well pay for itself in the first months/years of usage. I’ll take a look and see what I can find on those sites.
Edit: okay, trying to shop through google translate / currency calculator is actually aids so I’m gonna teach a man to fish instead. This is what I should have done from the start anyway.
Power supply: Anything from a decent brand, at basically anything >450W. a 650W or 850W is totally fine if it’s at a decent price. They only draw the power they need, they don’t just constantly pull 850W if the downstream components aren’t calling for it.
CPU: 12400 is a fine cpu for what you’re doing. You’ll transcode at 720p no problem, 1080p maybe a single stream in real-time. I wouldn’t bank on more than that. Only downsides here are the relatively shallow core counts if you ever expanded into other workloads. Without access to used xeon boards/cpus, it might be a reasonable choice though. What I would say is look for something older but with more cores/threads if you can. For example, a 10900 or even 10700k would probably be a better server cpu than a 12400.
Memory: DDR4 platforms are a great way to save money, as long as you aren’t planning on expanding to inferencing on cpu. Get as much as you can. 32-64gb of ddr4 should be dirt cheap, especially if you find a cheap motherboard with 4 memory sockets.
Motherboard: If you want this thing to be versatile, you want 2x pci-e slots. Old gaming full-sized ATX boards are the way to go here. 1 slot for an HBA, 1 slot for a GPU, and that should be all you need. Bonus for as many open sata sockets as possible. 6-8 is pretty typical on 10th-12th gen gaming ATX boards.
GPU: gpus will be much more efficient at transcoding than an igpu, especially from older intel CPUs. A 1050, 2060, 3050, basically anything from the 10-series onward has a decent nvenc encoder that would work well with plex/jellyfin. My goto is generally old workstation cards, I use a p620 myself and it handles a single 4k encode job no problem. I’m not sure if they’re viably purchasable anywhere in your area, but I’d definitely look out for a P620, P1000, or T400. Great value in those cards.
Drives/HBA: there are inexpensive LSI HBA cards to expand how many drives you can attach to a system if you need them, all you need is a spare pci-e slot and a place to physically mount the drives. The cheapest way to start here is to look for a motherboard with 4-6 sata slots and use those. Hardware raid is functionally dead these days in the real world, just use zfs or mdadm under linux to create an array with your desired level of resiliency/capacity.
Once you’ve priced out what it would cost to buy all of this new, look for prebuilt gaming PCs and office PCs that might be able to be expanded to fit these requirements. Prices look kind of steep on those markets you listed, but I’m sure something exists if you look hard enough.
I am also curious
Thank you for letting me know what software not to use; good bot
Crossfading and normalization would both independently be dealbreakers for me. I can’t go back
I would be genuinely surprised if fair use draws the line on format-shifted, legally purchased media, at “remote watch-together”, leaving format-shifting and local watch-together in-tact.
If it were up to the studio’s interpretation of the law, you’d need to purchase a license for each person during local watch-together.
RE: backups, I’d recommend altering your workflow. Instead of taking an image of a box, automate the creation of that box. Create a bash script that takes a base OS, and installs everything you use fresh. Then have it apply configuration files where appropriate, and lastly figure out which applications really need backup blobs to work properly (thunderbird, for example). Once you have that, your backups become just the data itself. Photos, documents, etc. Everything else is effectively ephemeral because it can be reproduced through automation.
Takes a lot less space, is a lot more portable. And much better in scenarios where something in your OS is broken or you get a new computer and want to replicate your setup.
agree in principal, but in practice:
parents who live across the state
plexamp for music
+1 to all of this.
For ~3 years I ran a Debian system off of a raid 1 of 2 USB drives. I didn’t have the spare drive bay slots in my cs24-ty and I didn’t have the room for an expander.
SanDisk apparently didn’t consider my use case “warranty-voiding” and were content to replace them whenever they failed. (I was honest during the first warranty inquiry about how they were used; I doubt you could get away with this with modern SanDisk though) I had a 3-year warranty on the drives, and checking my email, I replaced a total of 11 over the 3 year period. The first 7-8 were before I moved logging to a zfs dataset on the spinners, which helped a lot as those 7-8 failures were all in year 1 with the constant journaling, writing, and syncing of mostly logs.
TL/DR: great for testing if drivers and hardware work; don’t do this in production
When I started learning Linux at work, the game I played with myself was i’d install Debian stable minimal on my primary workstation and I would not reinstall it ever. No matter what happened, I would always fix it.
I learned to install the basic subsystems to get a GUI and audio, learned the fun of Nvidia drivers to get xinerama and hw decoding working. In retrospect it seems trivial but as a new learner it was challenging and rewarding.
At one point I was trying to do something, and a guide online suggested installing some repo and installing newer libraries. I did so, and a week later I did a dist-upgrade (because I didn’t know any better) and when I rebooted I was presented with a splash screen for “crunchbang” linux.
Figuring out how to get back to Debian without breaking everything probably taught me more about packages, package managers, filesystems, system config files, init (systemd wasn’t really a thing yet) than everything else I had done combined.
For anyone wondering: 12 years into the project I had a drive from the mdadm mirror die, and while mdadm was copying to another mirror, the other drive died. I considered that a win but y’all can be the judge (no files were lost, 12yr into my Linux journey I had long since figured out automating NFS and rsync).
There’s nothing magic about Soylent for weight loss. It’s a simple equation of calories in and calories out. The advantages that Soylent offered me was convenience for counting said calories, convenience for meal prep, and being reasonably certain my body was getting a decent distribution of micronutrients
Yes, 375 -> 250
I did ~1.5 years of only Soylent, then transitioned into 2/3 meals per day being Soylent, which I’ve done for the last ~6-7yrs.
I’m the healthiest I’ve ever been, but it does require discipline, exercise and attention like anything else. Calories are calories and if you consume more than you burn, you’ll poop a lot and gain weight. If you drink at a significant deficit (my 1.5years was at 1200kcal/day) you will poop once or twice a week and it will take a few months of your body getting used to it for it to be more than liquid.
As others have said though, it’s a deceptively dehydrating liquid. You absolutely still need to drink water, and your water intake will largely dictate how much you pee.
I do this with awesomewm. You define window startup behavior in the main config. Applications can have static behavior to start in certain places or will default to “wherever my cursor currently is”. I suspect i3 has similar functionality