Hacker Newsnew | past | comments | ask | show | jobs | submit | turbocon's commentslogin

I want to know if this is any different than all of the AMD AI Max PCs with 128gb of unified memory? The spec sheet say "128 GB LPDDR5x", so how is this better?

https://nvdam.widen.net/s/tlzm8smqjx/workstation-datasheet-d...


> AMD AI Max PCs with 128gb of unified memory? The spec sheet say "128 GB LPDDR5x", so how is this better?

Framework's AMD AI Max PCs also come with LPDDR5x-8000 memory: https://frame.work/desktop?tab=specs


The GPU is significantly faster and it has cuda, though I'm not sure where it'd fit in the market.

At the lower price points you have the AMD machines which are significantly cheaper, even though they're slower and with worse support. Then there's apple's with higher memory bandwidth and even the nvidia agx Thor is faster in GPU compute at the cost of worse CPU and networking, and at the 3-4K price point even a threadripper system becomes viable that can get significantly more memory


> The GPU is significantly faster and it has cuda,

But (non-batched) LLM processing is usually limited by memory bandwidth, isn't it? Any extra speed the GPU has is not used by current-day LLM inference.


I believe just inference is bandwidth limited, prompt processing and other tasks on the other hand needs the compute. As I understand it, the workstation is also as a whole focused on the local development process before readying things for the datacenters, not just running LLMs

CUDA.

....what? Why?

I mean it's interesting and I'm probably going to buy the sardines but why is this here?


this what you're looking for? https://antiqueradio.org/art/Sony8-301WTelevisionManual.pdf

Chatgpt (especially deep research) is pretty good about digging up stuff like this


Thanks for that - yes, I haven’t quite gotten on the “just use AI search for everything now” bandwagon, but of course it makes a lot of sense that it’d be in there somewhere.

Guess I’m gonna go to a local service place with this PDF and the TV and see what they can do. I’m filled with anticipation for the day that I can boot up a terminal on Sony’s first TV and include it in one of my exhibits.

I do retro computing exhibits, in case you were wondering why I have all this junk… ;)


Have you tried this model finetuned for a similar purpose by roblox https://www.josefprusa.com/articles/open-hardware-in-3d-prin...


Wow, what a nightmare of a non-deterministic bug introducing library.

Super fun idea though, I love the concept. But I’m getting the chills imagining the havoc this could cause


Didn't someone back in the day write a library that let you import an arbitrary Python function from Github by name only? It obviously was meant as a joke, but with AIcolytes everywhere you can't really tell anymore...


There's one that loads code out of the best matching SO answer automatically https://github.com/drathier/stack-overflow-import


Why not go further? Just expose a shell to the internet and let them do the coding work for you /s


It's not really something to be sarcastic about.

I've actually done this, setting aside a virtual machine specifically for the purpose, trying to move a step towards a full-blown AI agent.


Why on earth did you want to do that?


“Twitch does…”


Flask also started as an April 1st joke, in response to bottle.py but ever so slightly more sane. It gathered so much positive response, that mitsuhiko basically had to make it into a real thing, and later regretted the API choices (like global variables proxying per-request objects).


Is there somewhere I can read about those regrets?


Two days after the announcement: https://lucumr.pocoo.org/2010/4/3/april-1st-post-mortem/

I think there was another, later retrospective? Can't find it now.


I second this, I need to know more. programming lore is my jam.


It's like automatically copy-pasting code from StackOverflow, taken to the next level.


Are there any stable output large language models? Like stablediffusion does for image diffusion models.


If you use a deterministic sampling strategy for the next token (e.g., always output the token with the highest probability) then a traditional LLM should be deterministic on the same hardware/software stack.


Wouldn't seeding the RNG used to pick the next token be more configurable? How would changing the hardware/other software make a difference to what comes out of the model?


> Wouldn't seeding the RNG used to pick the next token be more configurable?

Sure, that would work.

> How would changing the hardware/other software make a difference to what comes out of the model?

Floating point arithmetic is not entirely consistent between different GPUs/TPUs/operating systems.


Deterministic is one thing, but stable to small perturbations in the input is another.


> Deterministic is one thing, but stable to small perturbations in the input is another.

Yes, and the one thing that was asked about was "deterministic" not "stable to small perturbations in the input.


This looks "fun" too: commit fixing a small typo -> the app broke.


So nothing's changed, then :D


Sounds like a fun way to learn effective debugging.


It imports the bugs as well. No human involvement needed. Automagically.


I mean, we're at the very early stages of code generation.

Like self-driving cars and human drivers, there will be a point in the future when LLM-generated code is less buggy than human-generated code.


That's a compiler with more steps.


Honestly if you haven't ever used got bisect I'd say you're missing out on a very powerful tool. To be able to, without any knowledge of the code base, isolate down to the exact commit that introduced a big is incredibly powerful


Everyone has there own preferences, but I'd look into uv if I were you. It allows you to specify the python version, and for scripts you can even specify the python version as part of the shebang


uv is literally the goat except I haven't able to make vllm work in uv for some reason. Though aside from that, I think I need to use shebang more because I don't use it as often right now.


Oooo entirely agree, I do suspect he'd come down on the "law is law" side however it's not clear to me a law was actually broken.


Seriously fantastic tool, I'm going to use the hell out of this, I appreciate you creating this so much! Let me know if there's a way to donate.


Much appreciated. The only donation I can think of is if you help spread the word. I learned a tough lesson this week that it is extraordinarily hard to market a product lol

P.S. If it wasn't for @dang putting it in the second-chance-pool, this post would have never been seen by more than 2 people.


I maybe the only person on here that had no idea this is a thing, but thank you this is incredible


Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: