I wouldn't worry about Lemmy having as many users as reddit in the short term. Success is not just a measure of userbase. A system just needs a critical mass, a minimum number of users, to be self-perpetuating. For a reddit post that has 10k comments, most normal people only read a few dozen comments anyways. You could have half the comments on that post, and frankly the quality might go up, not down. (That said, there are many communities below that minimum critical mass at the moment.)
Lemmy is now a real alternative. When reddit imploded Lemmy wasn't fully set up to take advantage of the exodus, so a lot of users came over to the fediverse and gave up right away. There were no phone apps, the user interface was rudimentary, and communities weren't yet alive. Next time reddit screws up in a high profile way, and they will screw up, the fediverse will be ready.
Lemmy has way more potential than reddit. Reddit's leadership has always been incompetent and slow at fixing problems. The fediverse has been very responsive to user feedback in comparison.
You can google site:reddit.com whatever
But if you google site:lemmy.world whatever then you're losing a significant amount of results. To get good results, you need to know which Lemmy instances is likely to have your answer, and with communities duplicated over different servers, that can be tough.
In the end I find I prefer this federation model, although I'm not sure although I'm a bit concerned about funding it if it scales up to the size of Reddit (same with Mastodon vs twitter).
Lemmy contents are replicated by federated servers, so you might find what you want by using site:lemmy.world or other big instances because they might also has replicated contents from other smaller instances.
This has more to do with how bad Google has gotten, such that you're forced to add restrictions like Reddit to get rid of SEO sites and get useful answers. A proper working search engine would show these (and any that are found in Lemmy) high up by default.
I'm sure the search problem will be solved somehow. Like all the content is on each instance so its just a case of it being accessible and indexed by google I guess?
I'm sure it's already being indexed by Google.
But people like to add site filters like site:Reddit.com or site:stackoverflow.com to prevent google from barfing up a bunch of garbage results on the front page, when they know that's probably where the results they want will be.
There is no way to add a Lemmy-wide filter to a Google search, because Lemmy instances are all different sites
Does it actually matter though because Lemmy contents are replicated by federated servers, thus big Lemmy instances such as lemmy.world might have contents from smaller federated instances as well. Try using site:lemmy.world next time and see if it'll improve the search result, though Lemmy.world is just 2 months old so maybe Google hasn't indexed it all
That's a good point.
If you filter by a major site, then it'll have content from all the major communities.
That won't help if you're looking for niche content, but that's not as important.
I wonder how replicated data shows up to the indexer. I don't know enough about search engine indexing or SEO. Will google index replicated data? Presumably it won't index feeds or searches, it'll index the actual posts, and I wonder if replicated posts are considered posts for the purposes of indexing or if the indexer will only look at local posts.
Google isn't thrilled with duplicate content. Following this thread here, it sounds like identical content might be hosted on multiple servers? If that is so, it's not going to be high value in Google's eyes.
If it's indexed, you'll be able to search it with Boolean modifiers, but it might not get priority in organic searches.
Yes, contents are replicated across federated instances. For example, here is the link to this thread on my instance: https://lemmy.institute/post/49173
If you check the html source there, there is a canonical link in the header that points to https://sh.itjust.works/post/2334723 , which is in the OP's instance. I think google will respect canonical links when indexing duplicated contents, so maybe the SEO aren't affected too much?
Presumably how it should work is that that even if content is duplicated, the crawlers would only index the "local" for Mastodon/Lemmy/etc servers, so they wouldn't see the duplication.
But idk how it actually works, and we're right back with my original concern of site filters
Ideally it would be popular enough that you wouldn't need the site modifier. Google would see that Lemmy has the most seen and perpetuated answer just like it sometimes does with Reddit now, whatever the instance.
People still often out the site modifier on just to prevent google from barfing up a bunch of crap they don't care about, even if they know that Reddit results will be near the top.
But once a site is popular enough for traffic and engagement to influence it's position in search, it's def going to be popular enough for bots, trolls, bad faith actors, grifters, etc.
Welcome to the old Internet. Decentralization is good in a way, people will have to try harder instead of having everything spoon fed to them by Google.