Why is the fact that she’s a us citizen in the last paragraph?
EDIT: I phrased this poorly, I meant, why hold that information back until the last paragraph? That’s sort of the key part of the story.
Why is the fact that she’s a us citizen in the last paragraph?
EDIT: I phrased this poorly, I meant, why hold that information back until the last paragraph? That’s sort of the key part of the story.
Ollama and openwebui for a nice web interface.
Besides the user engagement reasons, it’s pretty obvious why Elon Musk specifically would want people to not check sources. Nothing better for Xitter disinformation campaigns than not having readily available links to other sources.
What a Xithole.
But, I thought Trump was all about state’s rights? This is very confusing. /s
I’ve been tempted to ditch my current password manager and move to bitwarden. I think this is the final push I needed.
Bench warrant, let’s do this!
If you go, definitely stay at Four Seasons Total Landscaping next door, best accommodations around and their convention spaces are great for any press conferences you might need to hastily put together.
I used to think they were bots. I still do, but I used to, too.
First a caveat/warning - you’ll need a beefy GPU to run larger models, there are some smaller models that perform pretty well.
Adding a medium amount of extra information for you or anyone else that might want to get into running models locally
If you look at https://ollama.com/library?sort=featured you can see models
Model size is measured by parameter count. Generally higher parameter models are better (more “smart”, more accurate) but it’s very challenging/slow to run anything over 25b parameters on consumer GPUs. I tend to find 8-13b parameter models are a sort of sweet spot, the 1-4b parameter models are meant more for really low power devices, they’ll give you OK results for simple requests and summarizing, but they’re not going to wow you.
If you look at the ‘tags’ for the models listed below, you’ll see things like 8b-instruct-q8_0
or 8b-instruct-q4_0
. The q part refers to quantization, or shrinking/compressing a model and the number after that is roughly how aggressively it was compressed. Note the size of each tag and how the size reduces as the quantization gets more aggressive (smaller numbers). You can roughly think of this size number as “how much video ram do I need to run this model”. For me, I try to aim for q8 models, fp16 if they can run in my GPU. I wouldn’t try to use anything below q4 quantization, there seems to be a lot of quality loss below q4. Models can run partially or even fully on a CPU but that’s much slower. Ollama doesn’t yet support these new NPUs found in new laptops/processors, but work is happening there.
It’s a good thing that real open source models are getting good enough to compete with or exceed OpenAI.
I’ll preface by saying I think LLMs are useful and in the next couple years there will be some interesting new uses and existing ones getting streamlined…
But they’re just next word predictors. The best you could say about intelligence is that they have an impressive ability to encode knowledge in a pretty efficient way (the storage density, not the execution of the LLM), but there’s no logic or reasoning in their execution or interaction with them. It’s one of the reasons they’re so terrible at math.
I like the game, but agree with the over-tutorialed complaints. They have two difficulty modes, I wish only story mode got all the handholding. I think there’s enough obvious indicators to get you through all the game mechanics.
It has been on my list to figure out how to move to forgejo, need to do it soon before the migration process breaks or gets awful.
Coming from c# then typescript and nextjs, rye feels very intuitive and like a nice bridge / gateway drug into python.
surely he’ll be less of a twat then. right?
I guess it wasn’t bacon I hate for breakfast yesterday.
Why do you hate bacon, are you a windmill?
Yeah, I phrased that poorly, I was painfully aware of the Haitians eating cats BS, I meant this more like “why wait until the last paragraph to drop that key bit of information?” Clumsy writing / story structure at the very least.