A command-line program to download media on OnlyFans

onlyfans-scraper A command-line program to download media, like posts, and more from creators on OnlyFans. Installation You can install this program by entering the following in your console: pip install onlyfans-scraper Setup Before you can fully use it, you need to fill out some fields in a auth.json file. This file will be created for you when you run the program for the first time. These are the fields: { “auth”: { “app-token”: “33d57ade8c02dbc5a333db99ff9ae26a”, “sess”: “”, “auth_id”: “”, “auth_uniq_”: “”, […]

Read more

Fault tolerant, stealthy, distributed web crawling with Pyppeteer

crawler-cluster Distributed, Fault-Tolerant Web Crawling. Multi-process, multiple workers Client process queues tasks in Redis. Worker nodes pull tasks from Redis, execute task, and store results in Redis. Client process pulls results from Redis. Pros: Worker nodes can run on any machine. Add or remove worker nodes at runtime without disrupting the system. Achieves fault-tolerance through process isolation and monitoring. Workers are ran as systemd services, where each service is the smallest possible processing unit (either a single browser with a […]

Read more

A california coronavirus scrapers are written using Python

california-coronavirus-scrapers An experiment in open-sourcing the web scrapers that feed the Los Angeles Times’ California coronavirus tracker. The scrapers are written using Python and Jupyter notebooks, scheduled and run via GitHub Actions and then archived using git. Installation Clone the repository and install the Python dependencies. pipenv install Run all of the scraper commands. make Run one of the scraper commands. make -f vaccine-doses-on-hand/Makefile GitHub https://github.com/datadesk/california-coronavirus-scrapers    

Read more
1 2