Hacker Newsnew | past | comments | ask | show | jobs | submit | chairhairair's commentslogin

Already has complete support for SPM.


Anyone paying attention should have zero trust in what Sam Altman says.


What do you think his strategy is? He has to make money at some point.

I don’t buy the logic that he will “scam” his investors and run away at some point.


He makes money by convincing people to buy OpenAI stock.

If OpenAI goes down tomorrow, he will be just fine. His incentive is to sell the stock, not actually build and run a profitable business.

Look at Adam Neumann as an example of how to lose billions of investor dollars and still walk out of the ensuing crash with over a billion.

https://en.wikipedia.org/wiki/Adam_Neumann

His strategy is to sell OpenAI stock like it was Bitcoin in 2020, and if for some reason the market decides that maybe a company that loses large amounts of cash isn't actually a good investment... he'll be fine, he's had plenty of time to turn some of his stock into money :)


Why not build a profitable business like Zucc, Bill gates, Jensen, Sergey etc? These people are way richer much more powerful.


I believe, but have no proof, that the answer is "because it's easier to sell stock in an unprofitable business than build a profitable one", although given the other comment, there's a good chance I'm wrong about this :)


Altman doesn't have any stock. He's playing a game at a level people caught up on "capitalism bad" can't even conceptualize.


I'm more "capitalism good" (8 billion people on earth, 7 billion can read, 5 billion have internet, and almost no one dies in childbirth anymore in rich countries, which is several billion people), but that is really interesting that he has no stock and just gets salary.

I guess if other people buying stock in your company is what enables you to have a super high salary (+ benefits like company plane, etc), you are still kinda selling stock though, and honestly, having considered the "start a random software company aligned with the present trend (so ~2015 DevOps/Cloud, 2020 cryptocurrency/blockchain, 2024 AI/ML), pay myself a million dollar a year salary and close shop after 5 years because 'no market lol'" route to riches myself, I still wouldn't consider Altman to be completely free of perverse incentives here :)

Still, very glad you pointed that out, thanks for sharing that information ^^


Again incorrect. He doesn’t have a super high salary.


thanks for pointing it out ^_^


Holy shit you are right. He owns no equity and just gets a salary. I have no idea about the game he’s playing.


> He has to make money at some point.

Yes, but two paths to doing that are to a) build a profitable company, and b) accumulate personal wealth and walk away from a non-profitable company.

I'm not saying OpenAI is unprofitable, but nor do I see Altman as the sort who'd rule out option b.


I don't know how one would think doomers "ignore image and video AI models". They (Yudkowsky, Hinton, Kokotajlo, Scott Alexander) point at these things all the time.


It's completely apparent that HN dismissing doomers with strawmen is because these HN'ers simply don't even read their arguments and just handwave away based on vibes they heard through the grapevine


OpenAI simply can’t be trusted on any benchmarks: https://news.ycombinator.com/item?id=42761648


Remember that they've fired all whistleblowers that would admit to breaking the verbal agreement that they wouldn't train on the test data.


Could not find it on the open web. Do you have clues to search for?


Somewhat related, but I’ve been feeling as of late what can best be described as “benchmark fatigue”.

The latest models can score something like 70% on SWE-bench verified and yet it’s difficult to say what tangible impact this has on actual software development. Likewise, they absolutely crush humans at sport programming but are unreliable software engineers on their own.

What does it really mean that an LLM got gold on this year’s IMO? What if it means pretty much nothing at all besides the simple fact that this LLM is very, very good at IMO style problems?


Far as i can tell here, the actual advancement is in the methodology used to create a model tuned for this problem domain, and how efficient that method is. Theoretically then, making it easier to build other problem-domain-specific models.

That a highly tuned model designed to solve IMO problems can solve IMO problems is impressive, maybe, but yeah it doesn't really signal any specific utility otherwise.


I don't fault you for maintaining a healthy scepticism, but per the President of the IMO: "It is very exciting to see progress in the mathematical capabilities of AI models, but we would like to be clear that the IMO cannot validate the methods, including the amount of compute used or whether there was any human involvement, or whether the results can be reproduced. What we can say is that correct mathematical proofs, whether produced by the brightest students or AI models, are valid." [1]

The proofs are correct, and it's very unlikely that IMO problems were leaked ahead of time. So the options for cheating in this circumstance are that a) IMO are colluding with a few researchers at OpenAI for some reason, or b) @alexwei_ solved the problems himself - both seem pretty unlikely to me.

[1] https://imo2025.au/wp-content/uploads/2025/07/IMO-2025_Closi...


Is b) really that unlikely?


Not really. This whole thing looks like a deliberately planned PR campaign, similar to the o3 demo. OpenAI has enough talented mathematicians. They had enough time to just solve the problems themselves. Alternatively, some participants leaking the questions for a reward isn't very unlikely either, and I definitely wouldn't put it past OpenAI to try something like that. Afterwards, they could secretly give hints or tool access to the model, or simply forge the answers, or keep rerunning the model until it gave out the correct answer. We know from FrontierMath and ARC-AGI that OpenAI can't be trusted when it comes to benchmarks.


Yes


Why?


This is not a benchmark, really. It's an official test.


What is an _official_ test?


And what were the methods? How was the evaluation? They could be making it all up for all we know!


On OpenAI's own released papers they show Anthropic's models performing better than their own. They tend to be pretty transparent and reliable in honesty in their benchmarks.

The thing is, only leading AI companies and big tech have the money to fund these big benchmarks and run inference on them. As long as the benchmarks are somewhat publicly available and vetted by reputable scientists/mathematicians it seems reasonable to believe they're trustworthy.


Not to beat a dead horse or get into a debate, but to hopefully clarify the record:

- OpenAI denied training on FrontierMath, FrontierMath-derived data, or data targeting FrontierMath specifically

- The training data for o3 was frozen before OpenAI even downloaded FrontierMath

- The final o3 model was selected before OpenAI looked at o3's FrontierMath results

Primary source: https://x.com/__nmca__/status/1882563755806281986

You can of course accuse OpenAI of lying or being fraudulent, and if that's how you feel there's probably not much I can say to change your mind. One piece of evidence against this is that the primary source linked above no longer works at OpenAI, and hasn't chosen to blow the whistle on the supposed fraud. I work at OpenAI myself, training reasoning models and running evals, and I can vouch that I have no knowledge or hint of any cheating; if I did, I'd probably quit on the spot and absolutely wouldn't be writing this comment.

Totally fine not to take every company's word at face value, but imo this would be a weird conspiracy for OpenAI, with very high costs on reputation and morale.


The International Math Olympiad isn’t an AI benchmark.

It’s an annual human competition.


They didn’t actually compete.


Correct, they took the problems from the competition.

It’s not an AI benchmark generated for AI. It was targeted at humans



How many startups does this kill?


Japanese gaming industry employees have low economic stress and high feelings of freedom? I'd be surprised. Even if that is true, can that really be connected to housing availability? Again, I'd be surprised.


This in relation, of course, to the US/UK, where it has reached comical levels.


Yes, but please think of the poor shareholders..


If the current tech plateaus (but continues to come down in price, as expected) then this is a good prediction.

But, then there will be a demand for "all-in-one" reliable mega apps to replace everything else. These apps will usher in the megacorp reality William Gibson described.


>but continues to come down in price, as expected

Is it? AI very much seems to be in market capture mode. And IIRC, very few businesses actually report profits.

I can only predict AI models ramping up the cost like crazy once the victor captures the market. Same as every other tech trend in the last 20 years.


Hosting, bandwidth, storage, and compute all have come down by orders of magnitude in 20 years.

Regardless which model is currently the best, it looks like there will be an open weight model ~6 months behind it which can be vendored at costs that are closely tied to the hardware costs.


> Could you imagine someone saying, "be sure that the graphic for the molecule in figure 1 is 3D and has bright colors?"

Chemists are extremely brand-aware regarding their figures.

In synthetic chemistry many chemists could guess the author based just on the color scheme of the paper's figures.

For instance, look at the consistency here: https://macmillan.princeton.edu/publications/

And it comes with rewards! The above lab is synonymous with several popular techniques (one, organocatalysis, which garnered a Nobel prize) - the association would be much less strong if the lab hadn't kept a consistent brand over so many years.


Oh my, those figures are gorgeous! Thank you for sharing.


Altman: "Jony was running a design firm called LoveFrom that had established itself as really the densest collection of talent that I've ever heard of in one place AND HAS PROBABLY EVER EXISTED IN THE WORLD."

I felt physically sick from second-hand embarrassment watching this.


German has a word for second hand embarrassment. Fremdschämen. Comes in very handy here. If Sam continues like this it won’t be long until it becomes part of the regular English language like other German words such as Kindergarten.

And I’ll be happy that I don’t have to explain Fremdschämen anymore. Everything has its upsides.


> until it becomes part of the regular English language like other German words such as Kindergarten.

or Schadenfreude.


Somehow when the buzz-word machine found talent density, half the passengers forgot that density has a denominator. I see this goof a lot. If you accept the premise that Jony is literally the most talented human in the entire history of the world (I know I know), then obviously he was more dense sitting in a room alone, than after being diluted by hiring 50 other people.


Seems like a weird move to say:

* Jony Ive has built a company with the densest collection of talent in the world

* OpenAI is spending 10 figures to buy a company from Ive

* It is not the aforementioned company with the dense collection of talent; it's instead a company that no one has heard of


I almost had to tap out after the number Ive did on Altman in the early monolouge. Feels like some kind of office satire written by Mike Judge


Very trumpian language.


3b1b is a monetized partner?

Association with that brand would be very valuable.


Not yet! We don't monetize his content (it's not behind a paywall). But we are talking with him :)


Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: