

What part requires perfect precision?
If you want to parse sensor data, you do it in code before the LLM sees it.


What part requires perfect precision?
If you want to parse sensor data, you do it in code before the LLM sees it.


For sure, context rot is a problem, but that’s also the easiest thing to control for in this case. If sensor data is relevant to you, having some code to process and reduce it to a dashboard you can read is always a good idea, independently of getting an LLM into the loop.
This becomes more complicated with data you can’t really understand like results from blood tests, for example. But maybe you just don’t summarize any of that.


What we can call a health advisor is not a doctor. In fact, depending on the model, it will actively point you to seek medical help.


I can read minds and they’re thinking “we better get some money around here, otherwise we’re still blaming the immigrants”.
If they’re not great, it’s your fault /thread 😅


I believe right now it’s also valid to ditch NVIDIA given a certain budget. Let’s see what can be done with large unified memory and maybe things will be different by the end of the year.


chat is this socialism


somebody do something any day now


no more fokin ambushes


Trump has entered the chat


Let it die. A show this frequent can only end up being boring.


His whole existence is a financial demand. I believe Bloomberg calls this “a transactional period”. Put it plainly, y’all elected a corrupt president.


For some weird reason, in my country it’s easier to order a Beelink or a Framework than an HP. They will sell everything else, except what you want to buy.


Remind me of what are the downsides of possibly getting a framework desktop for christmas.


That’s a good point, but it seems that there are several ways to make models fit in smaller memory hardware. But there aren’t many options to compensate for not having the ML data types that allows NVIDIA to be like 8x faster sometimes.


For image generation, you don’t need that much memory. That’s the trade-off, I believe. Get NVIDIA with 16GB VRAM to run Flux and have something like 96GB of RAM for GPT OSS 120b. Or you give up on fast image generation and just do AMD Max+ 395 like you said or Apple Silicon.


I’m aware of it, seems cool. But I don’t think AMD fully supports the ML data types that can be used in diffusion and therefore it’s slower than NVIDIA.
it’s most likely math
You can even not have any data layer all together. The only thing missing from a local LLM is knowledge of current medications by name if you want to just say whatever prescription you’re following.