handlebar is written in Python and uses the Twitter API and Web Scraping. With handlebar, you can obtain the following information (if available) for all the members of the specified Twitter lists:
- Username
- Name
- Location
- Bio / Description
- Website URL
- Metadata description of above URL
You will need API access and Consumer access tokens to make calls to the Twitter API. These can be obtained by simply signing up on their developer page.
-
Clone the repo.
-
Install the python packages mentioned in
requirements.txt.- Using pip:
pip install -r requirements.txt
-
Edit the
settings.jsonfile.- In the
keyssection, fill out your Twitter API credentials.
{ "access_token": " #ACCESS TOKEN ", "access_secret": " #ACCESS SECRET ", "consumer_key": " #CONSUMER KEY ", "consumer_secret": " #CONSUMER SECRET " }- In the
all_urlssection, paste the URLs of the lists you would like to work on.
"all_urls": [ "https://twitter.com/abcd/lists/list1/", #Example list "https://twitter.com/xyz/lists/list2/members/lang=en" #Example list ]
- In the
formatsection, specify the extension the result files should be in.
"format": ".csv"
Currently only
.txtand.csvformats are supported- Set
count_only: trueif you wish to only count the total handles from all the lists. - Set
no_metadata: trueif you do not wish to scrape metadata from URLs obtained from user profiles. - Set
combine_files: trueif you wish to combine results from all the lists into one file called comibined_file.
- In the
Run the script from your terminal:
python3 main.py
The script will create a directory (if it doesn't already exist), called results and save the results of each list in a file corresponding to its name.
Example:
list1.csv list2.csv
- Atieve Wadhwa - AtieveWadhwa