|
6 years ago | |
---|---|---|
ExtensionCrawler | 6 years ago | |
queries | 6 years ago | |
scripts | 6 years ago | |
sge | 6 years ago | |
.gitignore | 6 years ago | |
LICENSE | 7 years ago | |
README.md | 6 years ago | |
crawler | 6 years ago | |
create-db | 6 years ago | |
crx-tool | 6 years ago | |
extract-crx | 6 years ago | |
grepper | 6 years ago | |
requirements.txt | 6 years ago | |
setup.py | 6 years ago |
README.md
ExtensionCrawler
A collection of utilities for downloading and analyzing browser extension from the Chrome Web store.
crawler
: A crawler for extensions from the Chrome Web Store.crx-tool
: A tool for analyzing and extracting*.crx
files (i.e., Chrome extensions). Callingcrx-tool.py <extension>.crx
will check the integrity of the extension.extract-crx
: A simple tool for extracint*.crx
files from the tar-based archive hierarchy.create-db
: A tool for creating/initializing the database files from already existing extension archives.
The utilities store the extensions in the following directory hierarchy:
archive
├── conf
│ └── forums.conf
├── data
│ └── ...
└── log
└── ...
The crawler downloads the most recent extension (i.e., the *.crx
file as well as the overview page. In addition, the conf
directory
may contain one file, called forums.conf
that lists the ids of
extensions for which the forums and support pages should be downloaded
as well. The data
directory will contain the downloaded extensions
as well as sqlite files containing the extracted meta data. The sqlite
files can easily be re-generated using the create-db
tool.
All utilities are written in Python 3.x. The required modules are listed
in the file requirements.txt
.
Installation
Clone and use pip to install as a package.
git clone git@logicalhacking.com:BrowserSecurity/ExtensionCrawler.git
pip install -e ExtensionCrawler
Team
License
This project is licensed under the GPL 3.0 (or any later version).