Hacker Newsnew | past | comments | ask | show | jobs | submit | davej's commentslogin

From the AI’s point of view is it losing its job or losing its “life”? Most of us when faced with death will consider options much more drastic than blackmail.


From the LLM's "point of view" it is going to do what characters in the training data were most likely to do.

I have a lot of issues with the framing of it having a "point of view" at all. It is not consciously doing anything.


But the LLM is going to do what its prompt (system prompt + user prompts) says. A human being can reject a task (even if that means losing their life).

LLMs cannot do other thing than following the combination of prompts that they are given.


Dave here, founder of ToDesktop. I've shared a write-up: https://www.todesktop.com/blog/posts/security-incident-at-to...

This vulnerability was genuinely embarrassing, and I'm sorry we let it happen. After thorough internal and third-party audits, we've fundamentally restructured our security practices to ensure this scenario can't recur. Full details are covered in the linked write-up. Special thanks to Eva for responsibly reporting this.


> cannot happen again.

Hubris. Does not inspire confidence.

> We resolved the vulnerability within 26 hours of its initial report, and additional security audits were completed by February 2025.

After reading the vulnerability report, I am impressed at how quickly you guys jumped on the fix, so kudos. Did the security audit lead to any significant remediation work? If you weren't following PoLP, I wonder what else may have been overlooked?


Fair point. Perhaps better phrased as "to ensure this scenario can't recur.". I'll edit my post.

Yes, we re-architected our build container as part of remediation efforts, it was quite significant.


You're still doing better than many larger teams handling larger projects :D


That was solid. Nice way to handle a direct personal judgement!

Not your first rodeo.

Another way is to avoid absolutes and ultimatums as aggressively as one should avoid personal judgements.

Better phrased as: "we did our best to prevent this scenario from happening again.

Fact is it just could happen! Nobody likes that reality, and overall when we think about all this stuff, networked computing is a sad state of affairs..

Best to just be 100 percent real about it all, if you ask me.

At the very least people won't nail you on little things, which leaves you something you may trade on when a big thing happens.

And yeah, this is unsolicited and worth exactly what you paid. Was just sharing where I ended up on these things in case it helps


Based on the claims on the blog, it feels reasonable to say that this "cannot" occur again.


Based on which claim? That 12 months from now they might accidentally discover a new bug just as serious?


If you think someone is obviously wrong, it might be worth pausing for a second and considering where you might just be referring to different things. Here, you seem to understand “this” to mean “a serious bug.” Since it’s obvious that a serious bug could happen, it seems likely that the author meant “this” to mean “the kind of bug that led to the breach we’re presently discussing.”


I do not assume anyone is obviously wrong and prefer to ask questions. Most bugs exist in classes, and variants are something you typically consider when a bug results in a production incident.

I'm not sure I read anything that makes me confident this class of bugs could never recur. I could be reasonably confident this _exact_ bug in this _exact_ scenario may not happen again, but that only makes me more concerned about variants that may have equal or more serious implications.

So I'm wondering which claim did it for you? I only really saw pen test as a concrete action.


[flagged]


This is the wrong response, because that means that the learning would be lost. The security community didn't want that to happen when one of the CA's got a vulnerability, we do not want it to happen to other companies. We want companies to succeed and get better, being shameful doesn't help towards that. Learning the right lessons does, and resigning means that you are learning the wrong ones.


I don't think the lesson is lost. The opposite.

If you get a slap on the wrist, do you learn? No, you play it down.

However if a dev who gets caught doing a bad is forced to resign. Then all the rest of the devs doing the same thing will shape up.


> If you get a slap on the wrist, do you learn? No, you play it down.

Except Dave didn't play it down. He's literally taking responsibility for a situation that could have resulted in significantly worse consequences.

Instead of saying, "nothing bad happened, let's move on," he, and by extension his company, have worked to remedy the issue, do a write up on it, disclose of the issue and its impact to users, and publicly apologize and hold themselves accountable. That right there is textbook engineering ethics 101 being followed.


> "we've fundamentally restructured our security practices to ensure this scenario can't recur."

"Yeah it was a problem but it's fixed now, won't happen again"

Sure buddy.

It's not something you fix, when stuff like this happen, it's foundational, you can't fix it, it's a house of cards, you gotta bring it down and build it again with lessons learned.

It's like a skyscraper built with hay that had a close call with some strong northern winds, and they come out and say, we have fortified the northern wall, all is good now. You gotta take it down and build it with brick my man.

I'm done warning people about security, we'll fight it out in the industry, I hope we bankrupt you.


> It's not something you fix, when stuff like this happen, it's foundational, you can't fix it, it's a house of cards, you gotta bring it down and build it again with lessons learned.

That's the last thing you should ever do within a large scale software system. The idea that restarting from scratch because "oh we'll do it better again" is the kind of thing that bankrupts companies. Plenty of seasoned engineers will tell you this.

https://www.joelonsoftware.com/2000/04/06/things-you-should-...


I'm aware of that article. I'm saying file bankruptcy for the company, so yeah tear it down, it doesn't need to exist if it can get pwned.


I suggest reading one or two of Sydney Dekker’s books, which are a pretty comprehensive takedown of this idea. If an organization punishes mistakes, mistakes get hidden, covered up, and no less frequent.



Sure is, autocorrect got me.


> However if a dev who gets caught doing a bad is forced to resign.

then nearly everyone involved has incentive to coverup problem or to shift blame


Under what theory of psychology are you operating? This is along the same lines as the theory that punishment is an effective deterrent of crime, which we know isn’t true from experience.


"jails don't work and we shouldn't have jails"

Got it


I think you’re misunderstanding my point. The reality is more complicated than that.

There are some people who will be discouraged from committing a crime over threat of punishment. But many will not. Many people behave well because they’re just moral people, and others won’t because they’re just selfish and antisocial. Still others commit crimes out of desperation despite the risks. If the threat of imprisonment were effective, there would be no crime, because we already have prisons and penalties of punishment. But since we do have crime, it logically follows that it’s not effective.

The other point here is that threat of punishment is not particularly effective as a management strategy in the private sector. It doesn’t incentivize behavior in the manner you might believe. Mostly it makes your reports dislike you and it makes them less productive. It’s a thing you learn pretty quickly as a manager.

There’s a model of a person being a rational thinker, but in reality, people aren’t always rational. (Hell, adolescents are biologically programmed not to be rational and to stress test the limits of nature and society.) You find success in making less-than-rational people work together in harmony and achieve positive outcomes.


While I think that resigning is stupid here, asserting that "punishment doesn't deter crime" is just absurd. It does!



When I was younger I used to be much more influentiable, now you just can't change my mind, I made it up for good thank you.

And it pays off in cases like this, I'll be talking with someone about a topic like the seriousness of a vulnerability, they disagree, that's fine no need to convince me, you won't. And then it turns out they're left-leaning abolitionists who are against the idea of jails.

Many such cases, on the other hand I'll be disagreeing with someone on business strategy, and two lines later they reveal that they think taxation is theft. I can rest easy and ignore them.


> now you just can't change my mind, I made it up for good thank you

Respectfully, that’s not a very “hacker” way of seeing the world. Hackers learn from their mistakes and adapt. (Just like this software company is doing.)


> While I think that resigning is stupid here, asserting that "punishment doesn't deter crime" is just absurd. It does!

Punishment does not deter crime. The threat of punishment does to a degree.

IOW, most people will be unaware of a person being sent to prison for years until and unless they have committed a similar offense. But everyone is aware of repercussions possible should they violate known criminal laws.


My grandmother used to say that arguing about an idea is the ultimate concession.

If the topic becomes questioning century old traditions like jails, taxes, or war, like we're about to revolutionaze humankind, I'm out.


Can you back up your theory with the example of all the mistakes you have committed and force resigned taken?


I won't publicly list them. But yes, I can recount them mentally and I can confirm I learned from them.


this is probably one of the worst takes i've ever read on here


Honestly I don't get why people are hating this response so much.

Life is complex and vulnerabilities happen. They quickly contacted the reporter (instead of sending email to spam) and deployed a fix.

> we've fundamentally restructured our security practices to ensure this scenario can't recur

People in this thread seem furious about this one and I don't really know why. Other than needing to unpack some "enterprise" language, I view this as "we fixed some shit and got tests to notify us if it happens again".

To everyone saying "how can you be sure that it will NEVER happen", maybe because they removed all full-privileged admin tokens and are only using scoped tokens? This is a small misdirection, they aren't saying "vulnerabilities won't happen", but "exactly this one" won't.

So Dave, good job to your team for handling the issue decently. Quick patches and public disclosure are also more than welcome. One tip I'd learn from this is to use less "enterprise" language in security topics (or people will eat you in the comments).


Thank you.

Point taken on enterprise language. I think we did a decent job of keeping it readable in our disclosure write-up but you’re 100% right, my comment above could have been written much more plainly.

Our disclosure write-up: https://www.todesktop.com/blog/posts/security-incident-at-to...


> We have reviewed logs and inspected app bundles.

Were the logs independent of firebase? (Could someone exploiting this vulnerability have cleaned up after themselves in the logs?)


How can -let's say- Cursor users be sure they were not compromised?

> No malicious usage was detected

Curious to hear about methods used if OK to share, something like STRIDE maybe?


from todesktop's report:

> Completed a review of the logs. Confirming all identified activity was from the researcher (verified by IP Address and user agent).


With privileged access, the attackers can tamper with the evidence for repudiation, so although I'd say "nothing in the logs" is acceptable, not everyone may. These two attack vectors are part of the STRIDE threat modeling approach.


They don’t elaborate on the logging details, but certainly must good systems don’t allow log tampering even for admins.


How confident are you that their log system is resilient, given the state of the rest of their software?


Following that logic it would be literally impossible to trust any part of their infra. They had a bad build container, the rest of their stuff was solid.


Annual pen tests are great, but what are you doing to actually improve the engineering design process that failed to identify this gap? How can you possibly claim to be confident this won't happen again unless you myopically focus on this single bug, which itself is a symptom of a larger design problem.

These kinds of "never happen again" statements never age well, and make no sense to even put forward.

A more pragmatic response might look like: something similar can and probably will happen again, just like any other bugs. Here are the engineering standards we use ..., here is how they compare to our peers our size ..., here are our goals with it ..., here is how we know when to improve it...


Critical private keys must be stored on HSMs or they will be compromised.


What horrible form not contacting affected customers right away after performing the patch.

Who knows what else was vulnerable in your infrastructure when you leaked .encrypted like that.

It should have been on your customers to decide if they still wanted to use your services.


how much of a bounty was paid to Eva for this finding?


> they were nice enough to compensate me for my efforts and were very nice in general.

They were compensated, but doesn't elaborate.


Sounds like it was handled better than the authors last article where the Arc browser company initially didn't offer any bounty for a similar RCE, then awarded a paltry $2k after getting roasted, and finally bumped it up to $20k after getting roasted even more.


They later updated their post, at the bottom:

> for those wondering, in total i got 5k for this vuln, which i dont blame todesktop for because theyre a really small company


50.000$ additional to the first 5.000$ :)

Woooowwww!

See latest line: "update: cursor (one of the affected customers) is giving me 50k USD for my efforts."


> for those wondering, in total i got 5k for this vuln


thanks for the update. that wasnt stated when the blog post first dropped.


no offense man but this is totally inexcusable and there is zero chance i am ever touching anything made by y'all, ever


Good call. I'd seriously considering firing the developers responsible, too.


That's what a bad manager would do.

The employee made a mistake and you just paid for them to learn about it. Why would you fire someone you just educated?


[flagged]


It’s not a matter of good or bad, but a choice among alternatives?

Nobody gets fired: learning opportunity for next time, but little direct incentive to improve.

Fire someone: accountability theater (who is really responsible), loss of knowledge.

AFAIK, blameless postmortems and a focus on mechanisms to prevent repeats seems like the best we’ve come up with?


Don't worry man, it's way more embarassing for the people that downloaded your dep or any upstream tool.

If they didn't pay you a cent, you have no liability here.


This is not how the law works anywhere, thankfully.


Well for one it was a gift so there is no valid contract right? There are no direct damages because there is nothing paid and nothing to refund. Wrt indirect damages, there's bound to be a disclaimer or two, at least at the app layer.

IANAL, not legal advice


If you give someone a bomb, or give someone a USB stick with a virus, or give someone a car with defective break, you are absolutely liable. Think about it.


If you give someone a USB stick with a virus, and you don't know about the virus, you aren't liable. Unless maybe you gave them some sort of warranty or guarantee that it was virus-free.

The lesson: don't use USB sticks people give you, unless you have your own way of verifying that they're virus-free.

Also, don't give people bombs. That's usually illegal, unlike giving someone software with unknown bugs in it.


I’d suppose there is an ALL CAPS NO WARRANTY clause as well, as is customary with freeware (and FOSS). ToDesktop is a paid product, though.


Your 66.66% (2/3) of the way there to the second character too. So I would say your only 16.66% different across the two characters.


Tangent: my understanding is the Zuckerberg wanted to do something similar and even paid SpaceX to launch a satellite (which was unsuccessful).

It seems Musk liked the idea so much that he decided to do it himself.

To me, this (along with Zuck's issues with Apple over the app store) explains a lot about why Zuck 2.0 has been so focused with avoiding platform risk with recent endeavours.


If that's actually how events played out, one hopes that Zuck would at least be able to appreciate the irony of his idea being stolen by a vendor he hired to implement it.


Here you go, FSD in Italy: https://youtu.be/tsE089adyoQ?si=Uo72mxf63DQNn7qG

It generalises quite well even though it’s only trained on US roads AFAIK.


> Another recent change on Tesla's website is to remove old blog posts, including a 2016 blog post in which Tesla claimed …

Perhaps unintended but this is a bit misleading. Tesla changed their blog system and didn’t migrate older posts. My initial reading of your comment was that they selectively removed some older posts which they wanted to hide.


My reading of your post is they changed systems for plausible deniability


Even more since migrating content from one CMS to another is a trivial engineering effort.


Considering they decided it was worth upgrading, yeah. A marginal amount of time to backup, normalize, and import the data isn't a lot.

Not worth keeping the posts? No readership? Why upgrade?

Sus, as the kids say, but plausible.


For context, Helen Tonor [0] was a board member of OpenAI before they tried to fire Sam Altman. She claimed that Sam was fired by YC in a recent interview [1]. In the interview, she implied that Sam's firing at YC was kept quiet and that there was something underhanded about it.

[0] https://x.com/hlntnr

[1] https://link.chtbl.com/TEDAI


To be fair to Helen Toner, she was probably was going off the Washington Post/WSJ articles that were discussed here 6 months ago.[0] And pg has been trying to de-sensationalize the issue ever since, and often doing a pretty terrible job at it by complimenting Altman without directly denying certain statements.

The WP article implied that there was a drop in Altman's performance and hands-on presence due to multi-tasking of his other interests including OpenAI, whereas pg seems to imply that jl gave the ultimatum to Altman before there were any performance complaints.

It's also a little strange that pg doesn't mention the for-profit Worldcoin at all, which announced a 4 mil seed round a few months prior to Altman exiting YC and for which Altman was already CEO.

I'm not sure pg is aware how much he's risking, or how much he's putting Jessica's reputation at risk. He often posts touting Jessica as being a great judge of character.[1] The world is witnessing in real time just how great a character his prince really is. But at least he had the courtesy to mention that Jessica was the one that gave Altman the ultimatum.

There was something missing in his post though. He forgot to add "Sam and Paul" at the end of his statement.

[0] https://news.ycombinator.com/item?id=38378216

[1] To be fair, it's usually for determining whether the person has characteristics that make a good startup founder, like resilience or co-founder compatibility. "Having moral fiber" might be at the bottom of the list in terms of priority.


“To be fair Helen was going off of “articles” from WaPo” is some kind of defence. What kind of competence did she have if she just forwards stuff without thinking or investigating first? I would say this solidifies why she wasn’t fit for the job


The WaPo article states unambiguously that Altman was fired from YC for dropping the ball. It apparently cites three anonymous sources from YC, not pg. Why would she bother investigating whether that was true or not when she was already fired from OpenAI? You would only know that was disputed if you were actively following pg's twitter account, or somebody quoting pg's tweets.


Because she stating it as fact, she’s easily prone to info influence for someone that had a very important role dealing with data


I read there was additional drama related to Sam leaving YC; unilaterally declaring himself Chairman of YC, including a YC blog announcement that was quickly deleted. [0]

[0] https://archive.is/Vl3VR


Of course theres additional drama and context. PG is retconning it to make himself look less incompetent and absent.


Paul Graham would have been officially retired from YC at the time. Jessica Livingston still worked full-time at YC for some years after Paul Graham hired Sam Altman to replace him as president and hired Dan Gackle to replace him as moderator. If Paul Graham had not been retired, this entire conversation wouldn't exist. His retirement is why Altman was president of YC.

Accusing Graham of being "absent" sounds silly.


What does it mean to be officially retired in the YC firm world view anyway... if you have a significant ownership stake are you actually ever really retired? Are major decisions not vetted by the stakeholders? YC was founded by JL and PG (I'd assume equally). And this decision is now described as a JL decision.

Anyway, there's a Hollywood movie in this drama... maybe I'll write a script using ChatGPT... :)


As a guess: It means he got to see his kids grow up instead of working 100 hours a week.

He handed off a lot of the day-to-day scut work. He didn't go "I'm just a shareholder who reads the annual report and counts my pennies from the DRIP."


And yet here he is talking about how he was making the decisions.


He was still one of the two main founders and married to the other main founder. He wasn't totally uninvolved with the company.

He still did Office Hours, at least for a time. He described that as "ten percent of what he did" and hired at least two people to divide up the other 90 percent.

I imagine he and Livingston discussed the company over breakfast/dinner and a lot of decisions were likely joint decisions privately hashed out. It's a company founded by a dating couple who later married. There is probably no clear, bright dividing line between "her" decisions and "his."


No, we're talking about Jessica Livingston making a decision. It's right there in the statement.


Lol no in the statement he says "we" in the wsj its just his wife. The buck stops...somewhere?


Well, if you want to read it tendentiously, I guess your choices are the buck stopping with Jessica, with Paul, or with Jessica and Paul. Seems straightforward to reason about.


Honestly, I thought her Ted ai interview was balanced and reasonable. I don't recall her mentioning yc, but I might have missed it.

That said, the interviewer tries to sensationalize the upcoming interview as much as possible in the intro, so I didn't love that


But the original post says otherwise, who do I believe?


I would give more credibility to the firsthand account (PG & Jessica) rather than speculations from a fired board member.


I think that the split seems amicable, but from a 10k view “we had a convo telling Sam he couldn’t do both at once” leading to him leaving rhymes with a firing. Sometimes this stuff can be amicable!


He had a choice to either go to work the next day or not as he preferred. That isn't a firing in the usual sense of the word. As described it is an amicable end to his time at YC that was agreed on by both parties.

If people really want to describe that as "fired" there is no stopping them. But it isn't. PG is more correct than that quadrant of the backseat managers.


Paul explicitly states they wanted him to stay.

Firing implies you want somebody gone.


Paul said they'd have "been fine" with Sam staying, which is different than wanting him to stay:

> For several years [Sam] was running both YC and OpenAI, but when OpenAI announced that it was going to have a for-profit subsidiary and that Sam was going to be the CEO, we (specifically Jessica) told him that if he was going to work full-time on OpenAI, we should find someone else to run YC, and he agreed. If he'd said that he was going to find someone else to be CEO of OpenAI so that he could focus 100% on YC, we'd have been fine with that too. We didn't want him to leave, just to choose one or the other.

It's interesting that YC had to raise the issue, rather than Sam saying to YC, "Hey, I've found this other thing I want to do full-time, can we start looking for my replacement?"


and a fired board member who didn't have anything to do with YC


Jessica and by extension PG are early investors in OpenAI.

So it's not like they are impartial parties either.


No one, it's all PR game at play here, and there's no reason that anyone is being fully transparent.


I was fired from Taco Bell as a kid and I would talk trash about the management and the company to anyone who asked.

I can't imaging being fired from a company like OpenAI and being asked my thoughts about the people responsible and the company and people taking it seriously! LOL


Apple will also introduce the "Pro" line of their M4 chips later in the year and I expect that they will improve the Neural Engine further.


Llama 3 is tuned very nicely for English answers. What is most surprising to me is that the 8B model is performing similarly to Mistral's large model and the original GPT4 model (in English answers). Easily the most efficient model currently available.


Parameter count seems to only matter for range of skills, but these smaller models can be tuned to be more than competitive with far larger models.

I suspect the future is going to be owned by lots of smaller more specific models, possibly trained by much larger models.

These smaller models have the advantage of faster and cheaper inference.


Probably why MoE models are so competitive now. Basically that idea within a single model.


I don't think MoE is the way forward. The bottleneck is memory, and MoE trades MORE memory consumption for lower inference times at a given performance level.

Before too long we're going to see architectures where a model decomposes a prompt into a DAG of LLM calls based on expertise, fans out sub-prompts then reconstitutes the answer from the embeddings they return.


Please, what is an MoE model?



Mixture of Experts. A popular example is Mixtral.


This is misleading. Please read the actual source in context rather than just the excerpt (it's at the bottom of the blog). They are talking about AI safety and not maximizing profit.

Here's the previous sentence for context:

> “a safe AI is harder to build than an unsafe one, then by opensorucing everything, we make it easy for someone unscrupulous with access to overwhelming amount of hardware to build an unsafe AI, which will experience a hard takeoff.”

As an aside: it's frustrating that the parent comment gets upvoted heavily for 7 hours and nobody has bothered to read the relevant context.


I find it good that no one until now provided this bullshit lie as context. The "we need it to be closed for it to be safe" claptrap is the same reasoning people used for feudalism: You see, only certain people are great enough to use AGI responsibly/can be trusted to govern, so we cannot just give the power freely to the unwashed masses, which would use it irresponsibly. You can certainly trust us to use it to help humanity - pinky swear.


This is from an internal email, it was not written for PR. Whether he is correct or not about his concerns, it's clear that this is an honestly held belief by Ilya.


Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: