Skip to content

Latest commit

 

History

History
120 lines (72 loc) · 4.21 KB

README.md

File metadata and controls

120 lines (72 loc) · 4.21 KB

publish-pages

Data Dumps: https://data.lemmyverse.net/

This project provides a simple way to explore Lemmy Instances and Communities.

List of Communities

The project consists of four modules:

  1. Crawler (NodeJS, Redis) /crawler
  2. Frontend (ReactJS, MUI Joy, TanStack) /frontend
  3. Deploy (Amazon CDK v2) /cdk
  4. Data Site (GitHub Pages) /pages

FAQ

Q: How can I set a link to automatically set the home instance?

You can append home_url and (optionally) home_type to the URL to set the home instance and type.

?home_url=lemmy.example.com ?home_url=kbin.example.com&home_type=kbin

home_type supports "lemmy" and "kbin" (default is "lemmy")

Q: How does discovery work?

It uses a seed list of communities and scans the equivalent of the /instances federation lists, and then creates jobs to scan each of those servers.

Additionally, instance tags and trust data is fetched from Fediseer.

Q: How does the NSFW filter work?

The NSFW filter is a client-side filter that filters out NSFW communities and instances from results by default. The "NSFW Toggle" checkbox has thress states that you can toggle through:

State Filter Value
Default Hide NSFW false
One Click Include NSFW null
Two Clicks NSFW Only true

When you try to switch to a non-sfw state, a popup will appear to confirm your choice. You can save your response in your browsers cache and it will be remembered.

Q: How long till my instance shows up?

How long it takes to discover a new instance can vary depending on if you post content that's picked up by one of these servers.

Since the crawler looks at lists of federated instances, we can't discover instances that aren't on those lists.

Additionally, the lists are cached for 24 hours, so it can take up to 24 hours for an instance to show up after it's been discovered till it shows up.

Q: Can I use your data in my app/website/project?

I do not own any of the data retrieved by the crawler, it is available from public endpoints on the source instances.

You are free to pull data from the GitHub pages site:

Lemmyverse Data Site

Please don't hotlink the files on the public website https://lemmyverse.net/

Q: How often is the data updated?

Currently, I upload a Redis dump generated by the crawler each night to s3, GitLab builds the JSON dump from that.

Data is also available from the artifacts of this action. You can also download Latest ZIP (using nightly.link)

dist-json-bundle.zip file contains the data in JSON format:

  • communities.full.json - list of all communities
  • instances.full.json - list of all instances
  • overview.json - metadata and counts

Crawler

Crawler README

Frontend

Frontend README

Data Site

Data Site README

Deploy

The deploy is an Amazon CDK v2 project that deploys the crawler and frontend to AWS.

config.example.json has the configuration for the deploy.

then run cdk deploy --all to deploy the frontend to AWS.

Similar Sites

Lemmy Stats Pages

Thanks / Related Lemmy Tools

Credits

Logo made by Andy Cuccaro (@andycuccaro) under the CC-BY-SA 4.0 license.