Hacker News new | past | comments | ask | show | jobs | submit | leodriesch's comments login

I got SAM 1 to work with MPS device on my MacBook Pro M1, don’t know if it works with this one too.


I ran it with 3040x3040px images on my MacBook M1 Pro in about 9 seconds + 200ms or so for the masking.


I’d guess testing hardware is same as training hardware, so A100. If it was on a mobile device they would have definitely said that.


I think it’s fair to leave it out in the on-device model comparison. 3b is much smaller than 8b, it is obviously not going to be as good as llama 3 if they did not make groundbreaking advancements with the technology.


The model is always wrong, since it predicts a propability distribution over all possible tokens, but the target has 100% possibility for one token and 0 for all others.


I am really impressed by the Apple Maps implementation. I think it also uses textured polygons, but does so in a very good looking way and at 120 fps on an iPhone, showing even a whole city in textured 3d.


Apple bought a Swedish startup called C3 and their became 3D part of Apple Maps. That startup was a spin-off from Saab Aerospace, who had developed a vision system for terrain-following missiles. Saab ran a project with the municipal innovation agency in Linköping and the result was that they decided this tech should be possible to find civilian use cases for. C3 decided to fly small Cessnas in grids across a few major cities and also Hoover Dam, and built a ton of code on top of the already extremely solid foundation from Saab. The timing was impeccable (now many years ago) and they managed to get Microsoft, Apple and Samsung into a bidding war which drove up the price. But it was worth it for Apple to have solid 3D in Apple Maps and the tech has stood the test of time.


I remember seeing a Nokia or Here demo around that time that looked like similar or the same tech. Do you know anything published about it with technical details? Seems like enough time has passed that it would be more accessible. I would love to learn more about it.


Brave Search also has one https://brave.com/search/api/


Better pricing than Bing, as well. And their summary feature is pretty good.


How does this compare to MLX? As far as I understand MLX is equivalent to PyTorch but optimized for Apple Silicon.

Is this meant for training MLX models in a distributed manner? Or what is its purpose?


It looks like MLX is a part of this initiative. https://github.com/apple/corenet lists "MLX examples" as one of the components being released in April.


As mentioned in the "mlx_examples/open_elm": "MLX is an Apple deep learning framework similar in spirit to PyTorch, which is optimized for Apple Silicon based hardware."


Just skimming the README it looks like it’s a layer above MLX. So looks like a framework around it to ease ML


It's a layer on top of PyTorch, and it has code to translate PyTorch models into MLX.


So, is CoreNet the equivalent of Keras, whereas MLX is the Jax/PyTorch equivalent?


Sounds reasonable. Apple writes the following about MLX: "The design of MLX is inspired by frameworks like NumPy, PyTorch, Jax, and ArrayFire."


Not quite. The closest equivalent would be something like fairseq. It's config (yaml) driven.


I'd say most are thinking of Midjourneys success in image generation when talking about this kind of progress.


I'm too.

But I still see no evidence that this keeps improving and not plateauing at some (current?) level.


I’d say talent? Outside of OpenAI no team has been able to release a model as capable as GPT-4, and I’m unsure if the CIA has been prioritizing LLM experts in their hiring.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: