this post was submitted on 19 Aug 2023
54 points (100.0% liked)
Lemmy
12544 readers
118 users here now
Everything about Lemmy; bugs, gripes, praises, and advocacy.
For discussion about the lemmy.ml instance, go to !meta@lemmy.ml.
founded 4 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
If they include links back to the original lemmy instance, it should help a bit. At least in the past the ranking algorithm liked seeing lots of links from other sites.
I'm going to need to brush up on my SEO skills. That's really the last thing I'm looking for.
Sadly no, it won't help in that way either (although that wasn't the intent).
The lifecycle of a webpage in a browser is something like:
When google and others crawl the web looking for data to include in search they generally only index content from step 1, so it would only see the parameters passed to LBS (shown as "declaration" in the demo), and would not see anything rendered by LBS.
This is the "static" nature of static sites. The page downloaded by the browser is not built on request, rather it's only built periodically by the author - usually whenever they have an update or another post. I haven't posted anything on my blog in months so the pages wouldn't have been re-built during that time. There are benefits to this strategy in that it's very secure, very robust, very fast, and very easy to host, but the disadvantage is that any dynamic or "up to date content" (like comments from lemmy) need to be prepared by the client and thus can not be included in step one above and indexed in search.
There is a best of both worlds approach (SSR) where you could render all the comments when a page is originally built, and then update it when the client later renders the page. This means there's at least something for search indexers to see even if it's not up to date. The problem here is that there's a plethora of different engines which people use to build pages and I can't make a plugin for all or even a few of them.
With all that in mind, this is fantastic feedback, and why I posted this pre-alpha demo. Lots of commenters have said the same thing. I can re-factor to at least make SSR easier.
I'll have to play with it. I haven't looked through the code, but I'm pretty sure the default Lemmy UI isn't SEO-friendly as well (though kbin probably is), so it would be cool to have a tool that provides it.
But maybe it doesn't make sense for it to be your tool. I think it would be cool to have a lemmy feature where it'll serve a cached, static page if it detects a bot is crawling it or the client doesn't support JavaScript, and serve the regular site otherwise. Maybe your tool is the right option, or maybe someone should build a different one. IDK.
You're right, this tool isn't designed to address this problem and is ill-suited.
Lemmy should definitely render a static page and then "hydrate" that with JavaScript in the client. This is a common problem with modern js apps. SSR (server side rendering) is the solution but it can be very complex. You really need to build the whole app with SSR in mind, it's not really something to bolt on as an additional feature at the end.
I think we could get away with sniffing the user agent and serving up static content instead if it looks like a bot.
I'm a full stack dev by day, but everything I've worked on has been a web app that doesn't need SEO, so I'm not exactly sure how that works in practice. But presumably we could generate and cache a basic webpage periodically for each post for use by bots and perhaps for accessibility (e.g. very basic HTML that's just good enough to use with links or something). It would have the same content as the main page, but none of the JS or CSS.
It shouldn't be too hard to render with a templating library.