[-] eddythompson@beehaw.org 12 points 1 year ago

How is it hard to use? It’s just a list of emojis by type with a search box like any phone.

[-] eddythompson@beehaw.org 3 points 1 year ago

How is it a money printing machine? They sell domains at cost based on ICANN fees. They don’t mark them up like other registrars, which is one of the main ways to make money in that business.

[-] eddythompson@beehaw.org 6 points 1 year ago

It clearly reads as autogenerated reply. It seems ambiguous to me still whether it’s thinking you’re trying to move your domains to squarespace and wondering if google sill keep data or if it’s about them moving domains to squarespace.

Though I’m general I’d assume if you move all your domains out of Google Domains before the transition, there shouldn’t be anything for them to transfer to squarespace.

[-] eddythompson@beehaw.org 11 points 1 year ago

What I remember attending a PHP event in ~2009 was one of the old veterans there saying:

Only Microsoft folks say "Sequel Server", we say "My S Q L"

[-] eddythompson@beehaw.org 2 points 1 year ago

wait, what??? how did my reply end up on this thread? did I screw up? I was replying to https://beehaw.org/post/506525 I think.

[-] eddythompson@beehaw.org 10 points 1 year ago

The search engine market isn't quite as diverse as it may appear https://www.searchenginemap.com/

There are maybe 4 or so 'crawlers', and the rest buys access to the part of their data they are willing to sell to others.

Running a crawler with the current size and complexity of the internet is expensive, and complicated. Then there is sifting and sorting the data in a reasonable searchable format, and then there is the quality problem, etc.

Much easier to license data access from a provider (Usually Bing or Google or both) and just offer some added features on top, like no tracking, different result UI, custom filtering values per Bing or Google's APIs that make your own "secret sauce", etc.

[-] eddythompson@beehaw.org 1 points 1 year ago

I think you sort of answer your own question when you wonder what the value of having tons of smaller instances versus a smaller number of larger ones as you discuss the stress of federation from the larger instances soon after.

I'm not sure I have. I was saying you must have/will end up with few large instances PERIOD. There is just no way around it. Otherwise you have a discovery/usability problem. Absolutely 0 non-techies will setup their own Lemmy instance. It's not even reasonable to assume they could. All I do for work is setup cloud deployments of applications and build tools to maintain and monitor them, and I have 0 interest in setting up a Lemmy or Mastodon instance. I like Mastodon in theory, if the main instance maintainers shit the bed, create an account on another instance. But I'm not setting up my own instance.

For example, I'm into woodworking, selfhosting, programming, and cooking. I'm not setting an instance for any of that. I just want to know what are the popular instances that have woodworking, programming, and cooking communities and just subscribe to them. I'm not gonna go looking for hundreds of small instances all over the internet, nor am I gonna host my own.

As for moderation, everything does not federate everywhere at once, there needs to be some interaction between the instances first (like subscriptions to communities) for things to really start transferring over. Yes, moderation will still take work, but you don’t have to address the entire fediverse at once.

Yes, that makes sense if the problematic aspects of another instance is just their content or culture. Like if there is a say ultra-religious instance, or a toxic incel instance, or a Nazi instance, etc. Sure, you never federate with them and you never have to worry about their culture.

This doesn't work for spammers/malicious actors instances though. As a spammer or malicious actor, all I have to do is setup dozens and dozens of random instances that are just constantly spamming every instance that's on a list somewhere. Forcing it to federate with my spam instances and putting load on them. As a bad actor, I actively come to you, not the other way around.

Then you're back to the allowlist, which breaks the "federation" idea in general or introduce a lot more mod/support cost needing to manually allow each instance that you "think" is not spam/malicious.

So finally in that a world of few inevitable large instances, and hundreds of thousands of small ones that are all polling/pushing activity to the big ones, that's just significantly more load on the system as a whole. There is no magic around cache, latency, inconsistency, slowness, etc except to be on the instance where you want to converse with people.

I tried googling for how ActivityPub scales to see if there is some write up and this came up for example https://lucumr.pocoo.org/2022/11/14/scaling-mastodon/

from that article:

One thing seems relatively certain: if Mastodon wants to host a sizable community where some people have followers from most other instances, then the size of an individual instance will matter a lot and I'm pretty sure that the only sensible approach will be to either not permit small instances to participate at all, or for those to come with some other restrictions that will require special handling.

Many developers don't want to accept the problem of back-pressure. (A topic I wrote about quite a bit incidentally). Unfortunately some bad servers can really break you, and you will have to avoid federating to them. In general too many small servers will likely cause issues for very popular accounts on popular servers.

[-] eddythompson@beehaw.org 2 points 1 year ago* (last edited 1 year ago)

well, there was a long thread about this on /r/selfhosted where @TheFrenchGhosty@lemmy.pussthecat.org @TheFrenchGhosty@libretooth.gr was saying pretty much what I said, but with a tad more mental gymnastics mostly about EU laws regarding reverse engineering and lack of a formal agreement between them and YouTube.

Unfortunately (or fortunately?), /r/selfhosted is private atm due to the blackout, so I'm unable to find and share thread link.

The facts are:

  • Invidious (as an OSS project) calls undocumented internal YouTube APIs (they call it InnerTube).
  • Anyone can host an Invidious instance.
  • The main Invidious instance, i.e: https://invidious.io/ received a cease and desist from YouTube.

@TheFrenchGhosty@lemmy.pussthecat.org @TheFrenchGhosty@libretooth.gr posted all about this on GitHub, reddit, their personal blog, and contacted random media outlets like the one linked here, to complain about how "we have nothing to do with YouTube, why is YouTube bullying us". And since everyone obviously wants to give the little guy the benefit of the doubt, everyone starts wondering how it could be that a project that's all about providing an alternative UI for YouTube, doesn't call YouTube.

It's like if a movie pirating website is trying to argue

"Endgame.mp4" is just a file name. It has nothing to do with Marvel or Disney. What the hell are those greedy companies have to do with us??

I'm all for invidious, piracy, etc. But seriously?

1

I was scrolling through a thread, and suddenly a green box popped up on the bottom left corner that said "Report created".

It looked clickable when I hovered over it, but nothing happened when I clicked it, then it disappeared.

What was that? was it something I did? I don't think I clicked anything. I was wondering if I clicked some 'report comment' by mistake, but the report comment button seems to be hidden under the 3 dot menu then it asks you for a bunch of extra stuff. So I definitely didn't click that by mistake

[-] eddythompson@beehaw.org 3 points 1 year ago* (last edited 1 year ago)

Because of course Invidious calls YouTube APIs. They call the internal APIs the same way YouTube official client calls the API. They even have the API Key of one of YouTube client's in their repo. The guy's argument is that since they reverse engineered the calls, which is fine, they don't have to agree to YouTube's TOS to call it, which means YouTube's cease and desist invalid. I host my own private instance of Invidious to stream youtube audio to my phone. Of course reverse engineering is fine, scarping is fine, even the code is fine, and I'd agree that YouTube going after repos on github is wrong. But of course hosting Invidious is a violation of YouTube's TOS.

[-] eddythompson@beehaw.org 3 points 1 year ago

PGP email has nothing to do with the email protocol. All your message metadata and headers are still not encrypted/can’t be encrypted. You can only encrypt some payload with a PGP key, and it’s up to the receiver to figure out whether or not they want to trust any of the message metadata. The entire envelope is still plaintext everywhere. PGP email is just email, but you’re sending some random encrypted text in it.

[-] eddythompson@beehaw.org 4 points 1 year ago

Not sure how anyone can say “GPG” and “Works great” in the same sentence tbh. GPG is a usability nightmare except for the most advanced users who use it. Good luck trying to get your house contractor or doctor or representative or non-techie friends and family or really anyone to give you their “public key”

[-] eddythompson@beehaw.org 5 points 1 year ago

I don’t disagree, but I think it’s a bit of an oversimplification to attribute it all to capital. There is a failure in how the original internet (and traditional FOSS for that matter) envisioned the world.

The original vision was that everything will be distributed. There are protocols, there are implementations, and there are “users”. Where the term “user” encapsulated everyone from the person developing/contributing/maintaining the code, the person deploying and operating it, all the way to the grandparent or child or otherwise absolutely non-technical end user.

The idea was sound. You are a technical user, you could run email server for a set of people you know. Others could do the same. Small companies could start offering paid services, etc.

But the devil is always in the details. Who is maintaining it? Who is keeping everything secure and updated? How does it scale? How frequently do you need to migrate everything because the operator is going out of business or has come down with health issues, or has died. How much trust do you have to put in every operator? People don’t want downtime. People don’t want frequent migrations. People don’t want to have to trust hundreds of small providers and have churn all the time in services they rely on for their day to day.

The rise of a centralized, large, and popular operators of each type of service is inevitable in that case. A couple of large email providers were always distant to happen. Same with storage, messaging, etc. It’s difficult to selfhost everything yourself, and it’s incredibly burdensome to do it for free for a large number of people.

view more: next ›

eddythompson

joined 1 year ago