![](https://lemmy.ca/pictrs/image/0a59bb89-bf0b-412c-b7b4-83247b4158bb.jpeg)
![](https://lemmy.world/pictrs/image/44bf11eb-4336-40eb-9778-e96fc5223124.png)
Yacht clubs often have Wednesday afternoon sailing races and those crews often need extra folks on board. I learned sailing that way when I was in university. An inexperienced reliable crew is member is better than an experienced unreliable crew.
Mastodon: @[email protected]
Yacht clubs often have Wednesday afternoon sailing races and those crews often need extra folks on board. I learned sailing that way when I was in university. An inexperienced reliable crew is member is better than an experienced unreliable crew.
Fun fact, DEDSEC is a type of memory used in Soviet era mainframes.
Ok, I’ll come clean, I’m actually an angry T-Rex that drives on the wrong side of the road. I’m just looking for validation
That chiropractic care is not evidence based
I’ll jump on board with this too. Someone’s got to be the smartest person on earth ¯\_(ツ)_/¯
If you’re driving the same speed as the car in front of you, you have no reason to use the left lane
What if the car in front of you is driving at the same speed but heading right at you? Or if there is an angry T-Rex in the right lane?
OK, I did the math. I could turn $1K into $74M in 10 years with that strategy. I lost a year of growth because I thought Ethereum was released in 2013 however it was the white paper that was released in 2013, the coin started public trading in August 2015. Either way, I’d still be happy with that return.
I’m doing this from memory to stay in the spirit of the question. I’d buy ethereum, then pivot to zoom in early 2020, pivot back to etherum mid 2020, then finally to NVIDIA at the end of 2022. I’ll look up how well my memory works have served me.
That’s 128GB RAM, the GPU has 24GB VRAM. Ollama has gotten pretty smart with resource allocation. Smaller models can fit soley on my VRAM but I can still run larger models on RAM.
I’ve installed Ollama on my Gaming Rig (RTX4090 with 128GB ram), M3 MacBook Pro, and M2 MacBook Air. I’m running Open WebUI on my server which can connect to multiple Ollama instances. Open WebUI has it’s own Ollama compatible API which I use for projects. I’ll only boot up my gaming rig if I need to use larger models, otherwise the M3 MacBook Pro can handle most tasks.
I’m running a search instance on a VPS so my home IP isn’t linked to my searches. The main disadvantage is that my VPS is in Toronto and I live 2hrs away so geo searches don’t work very well. For instance, if I Google “restaurants” I get results for local restaurants whereas if I Gregle (I named my search engine Gregle) I get results for results near my VPS.
DM me if you want a link to my instance to check it out. It’s open but I don’t publicize it because bad actors could ruin my IP addresses reputation with spam queries via the API.
That’s exactly what a virus that was trying to trick me would say…