Hacker Newsnew | past | comments | ask | show | jobs | submit | more ZacnyLos's commentslogin

> it's basically a carbon copy of twitter circa 2015, down to an almost identical UI. Except that there's no monetization, no ads, no growth hacking, which means that in the main features are there to serve the user

Wow, that's just like Mastodon.


BlueSky's big killer features that Mastodon fails at:

1) Better (optional) algorithmic feeds. Mastodon's "explore" is weaker than Bluesky's "popular with friends" and "discover"

2) Quote-tweets.

3) Easier onboarding. Mastodon forces you to care about which server you're on and it does matter and migrating later is hard. Meanwhile, BSky has "starter packs" that people can produce for each other with lists of users to follow to easily jump into a community.

4) Username-as-domain is better than the Mastodon "confirmed links in profile" thing for self-verified accounts.

I wish the properly-federated OSS community-funded one had won but I'll take either to be done with Twitter.


I don't think either platform is going to "win" in the sense of reaching the size and influence of Twitter, but both will hopefully be more resilient than Twitter was.


Yeah, "resilience" is my big concern. After Facebook and Twitter, I'm mostly concerned about a social network getting compromised by a moneyed political interest. Obviously I'm a liberal so I have my political opinion on the interest here, but the point stands for any alignment - I'd rather see something like Mastodon where people can fund and run their own servers that reflect their own values and interests, and then those servers can federate and defederate as is appropriate.

Basically: governance matters.


Fortunately using Bridgy Fed we can connect BlueSky with Fediverse and IndieWeb.

https://github.com/snarfed/bridgy-fed/


can I get a ELI10 on this - If I host a mastodon server, would I need to run this software along side it then and then everyone on my server could see these other platforms posts and such?


Accounts on any supported platform can follow the bot on their platform, and their posts get bridged to all the other ones as a subdomain of the bridgy bot. e.g. a mastodon instance with a handle for every bridged account.

It changed to require opting in to bridging because many mastodon users got very mad about it.


As long as this site doesn't implement the ActivityPub protocol, I don't see any reason for me to move to this site. I don't have time to maintain another account, and I want to keep in touch with people from Mastodon and Threads.


This feels like a very 2010s comment, when it was assumed that one has to be on every single social network app, copy/pasting the same 'content' into each of them, in order to be visible online. All that did was turn the net into a giant monoculture of hot takes and too-short posts. Only influencers and shovel-sellers need that.


I think it's more about the network effect: the point of social networks is to talk with your friends, so you want to be on a network your friends are on.


Being its own local place is a feature, IMO.


Twitter started irreversibly feeding users’ data into its “Grok” AI technology in May 2024, without ever informing them or asking for their consent.

https://noyb.eu/en/twitters-ai-plans-hit-9-more-gdpr-complai...


What does irreversibly mean in this context? It seems like negative connotations are implied, but I feel like it's like irreversibly baking a cake.


Once the data is "compressed" into the model it cannot be easily removed without starting the training over.


So you mean like

"He used one of my eggs to irreversibly make a cake"

It's true, but it would be kind of amazing if it weren't


Hmm, it's not that simple, is it? Let's say the AI is trained on the tweet "Ben Adams drove to Mexico yesterday but I still haven't heard from him."

From this knowledge, you can ask the AI "Who has driven to Mexico" and it might know that Ben Adams did, and reply with that.

HOWEVER it's also baked into the model and can't be surgically removed after a complaint. That's the irreversibility part. You can't undo isolated training. You need to provide it a new data set and train it all over again. They won't do that because it's too costly.

The problem with the above example is of course that it can also contain sensitive or private user details.

I've easily extracted the complete song lyrics to the letter from GPT-4 even if OpenAI try to put up guardrails against it due to the copyright issues. AI is really still in the wild west phase...


The irreversibility is still important to highlight, as it is distinctively different from a similar consent issue with search: "Google indexed my website against my will, but I will just forbid them to include me in search results going forward".


It is irreversible similar to how a student reading a textbook from LibGen can remember and profit from that information forever. Kinda crazy how many in this community went from champions of freedom of knowledge to champions of megacorps owning and controlling of all of human creation in the span of like two years when it became clear other corporations could profit off that freedom too.


More like

"He used his eyes to irreversibly read this post"


If they use Twitter data does grok answer with a 280 character text?

Additional Twitter data is in my eyes mostly low quality content, that's nothing I would want in a AI model.


> low quality content

How does it matter even if the quality is high or low? The point is user data was used without consent.


Yes, but nothing new, other AI models used data that they don't own. Makes it not better, but I think thats the path.


> If they use Twitter data does grok answer with a 280 character text?

That may be considered a feature.

ChatGPT seems reasonably concise, Gemini's answers tend to be verbose (without adding meaningful content).


I've lead myself to believe that long responses are actually beneficial for the quality of the responses, as processing and producing tokens are the only time when LLMs get to "think".

In particular, requesting an analysis of the problem first before jumping to conclusions can be more effective than just asking for the final answer directly.

However, this analysis phase, or similar one, could just be done hidden in the background, but I don't think any are doing that yet. From the user point of view that would be just waiting, and from API point of view those tokens would also cost. Might just as well entertain the user with the text it processes in the meanwhile.


My understanding is this used to be the case[1] but isn't really true any longer due to things like the "star" method for model training[2]. Empirically it absolutely (circa GPT3) used to be the case that if you prompted with "Explain all your reasoning step by step and then give the answer at the end" or similar it would give you a better answer for a complex question than if you said "Just give me the answer and nothing else" or similar, or asked for the answer first, and then circa gpt-4 answers started getting much longer even if you asked the model to be concise.

That doesn't seem to be the case any more and there has been speculation this is down to the star method being used for training newer models. I say speculation because I don't believe people have come out and said they are using star for training. OpenAI referred to Q* somewhere but they wouldn't be drawn on whether that * is this "star" and although google were involved in publishing the star paper they haven't said gemini uses it (I don't think).

[1] https://arxiv.org/abs/2201.11903

[2] https://arxiv.org/pdf/2203.14465


So did OpenAI, why is it only a problem when Twitter itself does it?


I'm pretty sure it's not. I'm pretty sure people have been angry about OpenAI doing the same thing for a while now.


Has it been proven that OpenAI used twitter for training? I know it knows about the popular tweets, but those are reported in many places, so could be ingested accidentally with other content.

(But regardless, many people raised an issue of OpenAI training from sources they shouldn't be allowed to access, so they're definitely a problem as well)


Twitter bad, but it s not unlawful in their jurisdiction . Don't want it? dont use it


As someone from the EU, hearing this argument over and over from Americans is exhausting.

They provide a product in the EU, therefore they must either follow EU law or exit the EU market. Just like an EU company that provides a product in the US has to follow US law.


I am in the EU.

The line of 'following the law of another country' is grey area on the internet, given that it goes both ways:

EU online companies providing services to US users fail to provide the free speech guarantees that the US laws afford their citizens. That's because all EU countries have more strict laws limiting free speech. Should the EU companies break their own countries' law to satisfy the US audience?


"EU online companies providing services to US users fail to provide the free speech guarantees that the US laws afford their citizens."

Exactly what is "free speech guarantees" in the context of a private business?


There are now states in the US which voted laws to regulate social media censorship. The US supreme court has declined ruling on them or taking them down based on companies' first amendment rights.

So it seems there are states where a europeans social medium should abide by rules that would most likely contradict european laws, right?


What are these state laws, can you give me an example?



> EU online companies providing services to US users fail to provide the free speech guarantees that the US laws afford their citizens. That's because all EU countries have more strict laws limiting free speech. Should the EU companies break their own countries' law to satisfy the US audience?

Could you sharpen up this claim? Like suppose I run a microblogging site but I delete libellous posts and incitements to violence in accordance with my local European law. Am I violating a US law by allowing Americans to use the site?


i m asking the same question


My understanding of your post was that you know that it violates US law and so you're asking what should be done. What I am asking is if it really does violate US law, and if so how.


I have a cochlear implant as well and can confirm it


So good post, thanks!


Mapy.cz


Interestingly both X (TT) and Instagram are trying to popularise their group functions, but it's not popular enough.


Bookwyrm — fediverse alternative: https://joinbookwyrm.com/


this is awesome! As soon as I saw this thread pop up, I thought, "this could be a fediverse app..."

It speaks to the value of fediverse that so many services are becoming fedi-ized. Not just Twitter, Instagram, Youtube, blogs, 4chan, reddit, but even GoodReads.

Thank you for sharing this link!


I have been using Bookwyrm for my personal reading list. Would recommend! Easy to set up.

I hear there are some issues with its licensing. They seem to use some sort of 'anti-capitalist' license which is non-OSI. But I think that its ok, we need to start experimenting again, and clearly the OSI model isn't working.


Do you recommend the instance you are on? I am looking to join, but the English speaking instance list was smallish, and I was having trouble choosing.


I am using https://bookwyrm.social (my username is 'davis' on it)

to be honest, I have not tried hard to use any of the social features. But hopefully soon I will try them and have more to say. But for now, all I need is the list of books I want to read, a way to track my progress, and also some light organization. This works great, and it's guilt-free compared to Goodreads!


Thank you!

I am especially excited that they have an import feature.


You might be interested in: https://indieblog.page/



Additional recommendation:

Ye Olde Blogroll: https://blogroll.org/


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: