Added SINGLE_PROCESS environment variable
This commit is contained in:
parent
b310ca2fc2
commit
2bcf6cbbf7
|
@ -2,3 +2,4 @@ RAPID7_API_KEY=
|
||||||
CACHE_SIZE=536870912
|
CACHE_SIZE=536870912
|
||||||
MASSDNS_HASHMAP_SIZE=1000
|
MASSDNS_HASHMAP_SIZE=1000
|
||||||
PROFILE=0
|
PROFILE=0
|
||||||
|
SINGLE_PROCESS=0
|
||||||
|
|
|
@ -49,7 +49,9 @@ Depending on the sources you'll be using to generate the list, you'll need to in
|
||||||
|
|
||||||
The so-called database (in the form of `blocking.p`) is a file storing all the matching entities (ASN, IPs, hostnames, zones…) and every entity leading to it.
|
The so-called database (in the form of `blocking.p`) is a file storing all the matching entities (ASN, IPs, hostnames, zones…) and every entity leading to it.
|
||||||
It exists because the list cannot be generated in one pass, as DNS redirections chain links do not have to be inputed in order.
|
It exists because the list cannot be generated in one pass, as DNS redirections chain links do not have to be inputed in order.
|
||||||
|
|
||||||
You can purge of old records the database by running `./prune.sh`.
|
You can purge of old records the database by running `./prune.sh`.
|
||||||
|
When you remove a source of data, remove its corresponding file in `last_updates` to fix the pruning process.
|
||||||
|
|
||||||
### Gather external sources
|
### Gather external sources
|
||||||
|
|
||||||
|
@ -143,7 +145,7 @@ This will download about 35 GiB of data the first time, but only the matching re
|
||||||
Note the download speed will most likely be limited by the database operation thoughput (a quick RAM will help).
|
Note the download speed will most likely be limited by the database operation thoughput (a quick RAM will help).
|
||||||
|
|
||||||
The script remembers which were the last sets downloaded, and will only import newer sets.
|
The script remembers which were the last sets downloaded, and will only import newer sets.
|
||||||
If you want to force re-importing, run `rm temp/rapid7_timestamps/sonar.*`.
|
If you want to force re-importing, run `rm last_updates/rapid7_*.txt`.
|
||||||
|
|
||||||
### Export the lists
|
### Export the lists
|
||||||
|
|
||||||
|
|
|
@ -53,7 +53,8 @@ function feed_rapid7 { # study, dataset
|
||||||
then
|
then
|
||||||
link="$(get_download_url $study $dataset)"
|
link="$(get_download_url $study $dataset)"
|
||||||
log "Reading $dataset dataset from $link ($old_ts -> $new_ts)…"
|
log "Reading $dataset dataset from $link ($old_ts -> $new_ts)…"
|
||||||
# curl -L "$link" | gunzip | ./feed_dns.py rapid7 $@
|
[ $SINGLE_PROCESS -eq 1 ] && EXTRA_ARGS="--single-process"
|
||||||
|
curl -L "$link" | gunzip | ./feed_dns.py rapid7 $@ $EXTRA_ARGS
|
||||||
if [ $? -eq 0 ]
|
if [ $? -eq 0 ]
|
||||||
then
|
then
|
||||||
echo $new_ts > $old_ts_file
|
echo $new_ts > $old_ts_file
|
||||||
|
|
|
@ -20,4 +20,5 @@ date +%s > "last_updates/massdns.txt"
|
||||||
massdns --output Snrql --retry REFUSED,SERVFAIL --hashmap-size "$MASSDNS_HASHMAP_SIZE" --resolvers temp/all_nameservers_ip4.list --outfile temp/all_resolved.txt temp/all_subdomains.list
|
massdns --output Snrql --retry REFUSED,SERVFAIL --hashmap-size "$MASSDNS_HASHMAP_SIZE" --resolvers temp/all_nameservers_ip4.list --outfile temp/all_resolved.txt temp/all_subdomains.list
|
||||||
|
|
||||||
log "Importing into database…"
|
log "Importing into database…"
|
||||||
pv temp/all_resolved.txt | ./feed_dns.py massdns
|
[ $SINGLE_PROCESS -eq 1 ] && EXTRA_ARGS="--single-process"
|
||||||
|
pv temp/all_resolved.txt | ./feed_dns.py massdns --ip4-cache "$CACHE_SIZE" $EXTRA_ARGS
|
||||||
|
|
Loading…
Reference in a new issue