One doesn’t need to panic buy items that they keep in stock at sufficient levels. :)
I am a Meat-Popsicle
One doesn’t need to panic buy items that they keep in stock at sufficient levels. :)
As I said, not for cost saving, but more for not needing to go out when people start panicking, or being stupid
Food, but not primarily for cost savings as most regular used things things don’t last longer than a year, which cost wise won’t bridge the gap.
55 lbs of 00 flour in the chest freezer, still have about 25lbs of AP flour in there. 1 30lb bag of Jasmine Rice, 1 25 lb basmati. I still have a ton of beans and and dry pasts in mylar/oxy absorb sitting in barens cans for long term storage. When covid started, I had 1 million calories in storage. I don’t plan to go back to that, but I intend to be able to hunker down for a long time.
For work, I’m pushing to purchase more laptops before tariffs.
I’ve considered stowing fuel with a stabilizer but even if prices double on fuel, I don’t use enough of it to make a difference.
It would be a good time to buy any lithium ion batteries and finish off those ali-express/temu orders.
Honestly, it had more validity behind it in the '80s. When you were just starting out your career you didn’t have a house yet you didn’t have any wealth amassed, The ideals of the left really shown through. But once you got older and started having some money, the fiscal Republican ticket sold you on tax cuts and provided boosts to help certain investment opportunities. It was still mostly just bullshit to make themselves rich but there was some financial opportunity there. It’s pretty much long gone for middle-class advantage anymore though.
You really need to get your backups in order.
I have a 5 digit slashdot id.
Not so much found out about but songs that didn’t used to bother me now kind of bother me. I was a very big Stone Temple Pilots fan, Even though the rhythms slap the songs are a little too rapey these days for my taste.
courts of law.
There are always courts of law! Just somtimes it’s the king in his court making up the law…
kbin obviously!
Minimum open services is indeed best practice but be careful about making statements that the attack surface is relegated to open inbound ports.
Even Enterprise gear gets hit every now and then with a vulnerability that’s able to bypass closed port blocking from the outside. Cisco had some nasty ones where you could DDOS a firewall to the point the rules engine would let things through. It’s rare but things like that do happen.
You can also have vulnerabilities with clients/services inside your network. Somebody gets someone in your family to click on something or someone slips a mickey inside one of your container updates, all of a sudden you have a rat on the inside. Hell even baby monitors are a liability these days.
I wish all the home hardware was better at zero trust. Keeping crap in isolation networks and setting up firewalls between your garden and your clients can either be prudent or overkill depending on your situation. Personally I think it’s best for stuff that touches the web to only be allowed a minimum amount of network access to internal devices. Keep that Plex server isolated from your document store if you can.
that includes a large 285MB ZIP archive to install a Linux VM with a pre-installed backdoor.
When I just use one of the older distros that fits on a floppy?
Yeah, a company got toasted because one of their admins was running Plex and had tautulli installed and opened to the outside figuring it was read-only and safe.
Zero day bug in tat exposed his Plex token. They then used another vulnerability in Plex to remote code execute. He was self-hosting a GitHub copy of all the company’s code.
We’re a long way from trusting it to do something critical without intervention.
AI would be good at looking at an X-ray after a doctor and pointing out anomalies. But it would be bad to have it tell the doctor that everything looks fine.
We have high standards for American Chinese food. There was this place where we used to live in the food was great. Not everything they made came out of a bag, and even the things that did come out of a bag had absolutely superior sauces. I don’t know exactly what they did but whatever it was it was better heads and tails than anything else around here.
We ordered our regular dishes one day. A few hours later we were exploding out of both ends. Was it them? was lunch? Who knows? We went about our regular business and two weeks later ordered the same regiment. A few hours later we again were exploding out of both ends.
The puking wasn’t all that bad but the raw acid diarrhea and the massive cramps were just insane.
This was a pretty bad scenario because of the time we lived in a house with one bathroom.
We never ordered from there again. They had this really great iced tea It took me ages to figure out how to replicate it. It ended up being like 14 to 1 regular sweetened black tea to Earl Gray, plus a splash of lemon.
I keep a root folder. On Windows it’s in c:\something on Linux it’s in /something
Under there I’ve got projects organized by language. This helps me organize nix shells and venvs.
Syncthing keeps the code bases and synced between multiple computers
I don’t separate work from home because they don’t live in the same realm.
Only home stuff in the syncthing.
It tells me what document in the collection it used, But it doesn’t give me too much in the way of context or anything about the exact location in the document. It will usually give me some wording if I’m missing it and I can go to the document and search for that wording.
I’m just one person searching a handful of documents so the sample size is pretty small for repeatability, so far, if it says it’s in there, it’s in there. It definitely misses things though, I’m still early in the process. I need to try some different models and perhaps clean up the data a little bit for some of the stuff.
Using the documentation as source data It doesn’t seem to hallucinate or insist things are wrong, it’s more likely to say I don’t see any information about that when the data is clearly in the data set somewhere.
YW on the responses I’m having fun with it even if it’s taking forever to get it to dial in and be truly useful.
Trident VGA?
I got a 3DFX voodoo as soon as they came out. GL quake was mind-blowing.
I bought a Riva TNT
Then a GeForce 2
Then a Radeon 9000
Then for a bunch of years I just moved into laptop after laptop with discrete GPUs.
Now I still have a 1080 and a 2070 doing a little bit of light AI work and video transcoding for me. But I’m still relying on crappy laptop GPUs for all my gaming. They’re good enough.
I have two projects for it right now. The first is shoving my labyrinth of HOA documents into it so I can answer quick questions about the HOA docs or at least find the right answer more effectively.
The second is for work, I shoved a couple months of slack, some Google docs, some PDFs all about our production product. Next I’m going to start shoving some of GitHub in there. It would be kind of nice to have something that I could ask where is the shorting algorithm and how does it work and it could give me back where the source code is in any documentation related to it.
The HOA docs I could feed into GPT, I’m out till you sleep apprehensive to handover all of her production code to a public AI though.
I’ve got it running on a 2070 super and I’ve got another instance running on a fairly new ARC. It’s not fast, But it’s also not miserable. I’m running on the medium sized models I only have so much VRAM to deal with. It’s kind of like trying to read the output off a dot matrix printer.
The natural language aspect is better than trying to shove it into a conventional search engine, say I don’t know what a particular function is called or some aspect or what the subcompany my HOA uses to review architectural requests. Especially for the work stuff when there’s so many different types of documents lying around. I still need to try some different models though my current model is a little dumb about context. I’m also having a little trouble with technical documentation that doesn’t have a lot of English fluff. It’s like I need it to digest a dictionary to go along with the documents.
I got olama and WebUI working privately / locally and I’m able to insert documents into it with persistence and query them.
PShaw, that’s how I had to do it. Slackware on floppy. Pre-internet search engine, one computer per household. No cellular data.
windows -> Dial up -> look at some docs, take nodes -> reboot into Slackware -> mess with the console -> get stuck -> reboot into windows -> repeat