advertools
Online marketing productivity and analysis tools
Crawl websites, Generate keywords for SEM campaigns, create text ads on a large scale, analyze multiple SERPs at once, gain insights from large social media posts, and get productive as an online marketer.
If these are things you are interested in, then this package might make your life a little easier.
To install advertools, run the following from the command line:
pip install advertools
# OR:
pip3 install advertools
- robots.txt
- XML Sitemaps
- SEO Spider / Crawler
- Crawl Strategies
- How to crawl a list of pages, and those pages only (list mode)?
- How can I crawl a website including its sub-domains?
- How can I save a copy of the logs of my crawl for auditing them later?
- How can I automatically stop my crawl based on a certain condition?
- How can I (dis)obey robots.txt rules?
- How do I set my User-agent while crawling?
- How can I control the number of concurrent requests while crawling?
- How can I slow down the crawling so I don't hit the websites' servers too hard?
- How can I set multiple settings to the same crawl job?
- I want to crawl a list of pages, follow links from those pages, but only to a certain specified depth
- How do I pause/resume crawling, while making sure I don't crawl the same page twice?
- How do I use a proxy while crawling?
- How can I change the default request headers?
- XPath expressions for custom extraction
- User-agent strings for use in crawling
- Crawl Analytics
- Crawl headers (HEAD method only)
- Crawl images
- Log File Analysis
- Parse and Analyze Crawl Logs in a Dataframe
- Reverse DNS Lookup
- Analyze Search Engine Results (SERPs)
- Google's Knowledge Graph
Indices and tables
- Index & Change Log
- advertools package
- Subpackages
- advertools.cli_docs package
- Submodules
- advertools.cli_docs.cli module
- Convert a robots.txt file (or list of file URLs) to a table in a CSV format
- Download, parse, and save an XML sitemap to a table in a CSV file
- Split a list of URLs into their components: scheme, netloc, path, query, etc.
- Crawl a list of known URLs using the HEAD method
- Parse, compress and convert a log file to a DataFrame in the .parquet format
- Perform a reverse DNS lookup on a list of IP addresses
- Generate a table of SEM keywords by supplying a list of products and a list of intent words
- Get stopwords of the selected language
- Get word counts of a text list optionally weighted by a number list
- Search for emoji using a regex
- Extract structured entities from a text list; emoji, hashtags, mentions
- Tokenize documents (phrases, keywords, tweets, etc.) into tokens of the desired length
- SEO crawler
- advertools.cli_docs.cli module
- Module contents
- Submodules
- advertools.code_recipes package
- Submodules
- 🕷 SEO Crawling & Scraping: Strategies & Recipes
- How to crawl a list of pages, and those pages only (list mode)?
- How can I crawl a website including its sub-domains?
- How can I save a copy of the logs of my crawl for auditing them later?
- How can I automatically stop my crawl based on a certain condition?
- How can I (dis)obey robots.txt rules?
- How do I set my User-agent while crawling?
- How can I control the number of concurrent requests while crawling?
- How can I slow down the crawling so I don't hit the websites' servers too hard?
- How can I set multiple settings to the same crawl job?
- I want to crawl a list of pages, follow links from those pages, but only to a certain specified depth
- How do I pause/resume crawling, while making sure I don't crawl the same page twice?
- How do I use a proxy while crawling?
- How can I change the default request headers?
- XPath expressions for custom extraction
- User-agent strings for use in crawling
- 🕷 SEO Crawling & Scraping: Strategies & Recipes
- Module contents
- Submodules
- advertools.cli_docs package
- Submodules
- Create Ads on a Large Scale
- Create Ads Using Long Descriptive Text (top-down approach)
- advertools.cli module
- Crawling and Scraping Analysis
- Emoji: Extract, Analyze, and Get Insights
- Extract structured entities from text lists
- 🕷 Python Status Code Checker with Response Headers
- Image Crawler and Downloader
- Import and Analyze Knowledge Graph Results on a Large Scale
- Generate Keywords for SEM Campaigns
- Log File Analysis
- Parse and Analyze Crawl Logs in a Dataframe
- Regular Expressions for Extracting Structured Entities
- Reverse DNS Lookup in Bulk
- 🤖 Analyze and Test robots.txt Files on a Large Scale
- Import Search Engine Results Pages (SERPs) for Google and YouTube
- Download, Parse, and Analyze XML Sitemaps
- 🕷 Python SEO Crawler / Spider
- Stopwords in Several Languages
- Twitter Data API
- URL Builders
- Split, Parse, and Analyze URL Structure
- Text Analysis
- Tokenize Words (N-grams)
- YouTube Data API
- Module contents
- Subpackages
- Change Log - advertools
- 0.16.4 (2024-12-19)
- 0.16.3 (2024-12-17)
- 0.16.2 (2024-12-08)
- 0.16.1 (2024-08-19)
- 0.16.0 (2024-08-18)
- 0.15.1 (2024-07-16)
- 0.15.0 (2024-07-15)
- 0.14.4 (2024-07-13)
- 0.14.3 (2024-06-27)
- 0.14.2 (2024-02-24)
- 0.14.1 (2024-02-21)
- 0.14.0 (2024-02-18)
- 0.13.5 (2023-08-22)
- 0.13.4 (2023-07-26)
- 0.13.3 (2023-06-27)
- 0.13.2 (2022-09-30)
- 0.13.1 (2022-05-11)
- 0.13.0 (2022-02-10)
- 0.12.3 (2021-11-27)
- 0.12.0,1,2 (2021-11-27)
- 0.11.1 (2021-04-09)
- 0.11.0 (2021-03-31)
- 0.10.7 (2020-09-18)
- 0.10.6 (2020-06-30)
- 0.10.5 (2020-06-14)
- 0.10.4 (2020-06-07)
- 0.10.3 (2020-06-03)
- 0.10.2 (2020-05-25)
- 0.10.1 (2020-05-23)
- 0.10.0 (2020-05-21)
- 0.9.1 (2020-05-19)
- 0.9.0 (2020-04-03)
- 0.8.1 (2020-02-08)
- 0.8.0 (2020-02-02)
- 0.7.3 (2019-04-17)
- 0.7.2 (2019-03-29)
- 0.7.1 (2019-03-26)
- 0.7.0 (2019-03-26)
- 0.6.0 (2019-02-11)
- 0.5.3 (2019-01-31)
- 0.5.2 (2018-12-01)
- 0.5.1 (2018-11-06)
- 0.5.0 (2018-11-04)
- 0.4.1 (2018-10-13)
- 0.4.0 (2018-10-08)
- 0.3.0 (2018-08-14)
- 0.2.0 (2018-07-06)
- 0.1.0 (2018-07-02)
- advertools package