Yet another goofy project for the SS13 server Lifeweb. This node.js project scans the Lifeweb discord for showlad entries and files, it then parses them down into usable data to generate a rough estimate of the amount of times a role has been played as well as providing user statistics. The final output will be a csv that is an amalgamation of both user posts and scraped text files containing much of the same data. The csv will be formatted for SQL or excel importing, so feel free to use whatever you like.
It is very likely that I will personally use this data for some APIs or to populate a DB, what you do with it is up to you.
This is not the web interface, this is just the data scraping and parsing tool.
Works in tandem with the web interface: https://github.com/Daelso/RavenStats
This project would not be possible without the DiscordChatExporter (It is seriously stupid hard to automate scraping discord channels, bless this dudes work) - https://github.com/Tyrrrz/DiscordChatExporter
Player data is collected via the Showlads commanda at the end of the round and the formatting and provision of it is spotty at best. Data collected is likely of horrific integrity and contains duplicates as well does not include pre-discord games nor games that no one bothered to post.
You will need node, npm and dotnet to run this program. A discord account with access to the Lifeweb Hideout is also required.
Clone down the repo locally and within your RavenStats folder create a file titled .env, this is where you will put your sensitive data.
ENV Format:
DISCORD_TOKEN = yourtokenhere
FILE_DIR = "C:/Users/yournamehere/OneDrive/Desktop/coding projects/RavenStats/showlads_dump"
To get your discord token, please view the instructions for the DiscordChatExporter here: https://github.com/Tyrrrz/DiscordChatExporter/blob/master/.docs/Token-and-IDs.md#how-to-get-user-token
File_dir refers to where you would like to export your contents to.
npm install
We use several dependencies so please install them before attempting to run the program.
Firstly, follow the installation steps above. Once everything is installed, you have a few scripts to choose from:
npm run full-scrape
This will run a complete scrape of general, questions, coolstorium and showlads populating your output folder with the files but not parsing them. Be aware this could take over 24 hrs.
npm run full-scrape-parse
This will run a complete scrape of general, questions, coolstorium and showlads populating your output folder with the files AND parsing them/cleaning up.
npm run parse
This will parse any of the files you have sitting in your output folder. Useful if you wanted to modify/add additional sourced data like pastebins manually after scraping or just wish to bundle up a few weeks worth.
npm run weekly-scrape
This is intended to be run on Mondays, it will scrape from the previous Friday until Monday, effectively giving you that weekends showlads.
npm run weekly-scrape-parse
Scraped like above, calls the parse function immediately so your output is only the formatted CSV