Does it generate only username1, username2 etc. or makes combo list username1:username1 ?
Okay, anyone knows how to convert these usernames into combos?
Okay, anyone knows how to convert these usernames into combos?
hey so i did all the steps but nothing happens in the api scrapper. it just stays there and does nothing for the past hour lolQuote:
[Only registered and activated users can see links. Click Here To Register...]
[Only registered and activated users can see links. Click Here To Register...]
How to get the API Keys:
0. Recommended to get a minimum of 10 API Keys
1. Get a lot of accounts, doesn't matter the level, they can be even level 1
2. Go to [Only registered and activated users can see links. Click Here To Register...] , login and get the API Key
3. You will need to login in this page with each account that you got, each account have an API key.
4. Insert the API keys in the ApiKeys.txt file, 1 api key per line.
5. Done.
In the API Scraper choose range between 1 and 90,000,000 the higher the number the newer the account (with high numbers the process might last a bit so dont worry if it doesnt scrape immediately)
[Only registered and activated users can see links. Click Here To Register...] #!MY5mHYzb!SslSY1wO9TQJ5UFoZIrCmPu-7VxkzIVEpzznopsYFYI
[Only registered and activated users can see links. Click Here To Register...]
I have a combolist converter that supports password list with a variable for usernames, so you can make a list like..Quote:
Does it generate only username1, username2 etc. or makes combo list username1:username1 ?
Okay, anyone knows how to convert these usernames into combos?
username:username username:username1 username:username12 username:username123 username:password1 username:abc123
1) I did not code thisQuote:
Just tried it out. Works a lot better than the janky API scraper I made in python does :p Also I like the addition to chose inactive accounts.
Few questions/suggestes:
1.) The naming conventions for the output texts are a bit confusing, especially if you are scraping multiple regions. Would it be possible to make it do like NA 1-10000.txt if you scraped 1 - 10000 on the NA server? etc. It seems like, currently, a scrape on NA from 1-10000 would be put in a text file named "ScrapeResult - 1.txt"
2.) When it's done running will it close automatically or should I manually close it?
3.) What is the criteria to mark an account inactive?