Bulk download papers from scihub for text mining

https://janikvonrotz.ch/2020/05/07/bulk-download-papers-from-scihub-for-text-mining/

Last evening I wanted to play some games with my brother. Instead we ended up writing a simple script to bulk download papers from scihub. He studies bioinformatics and is currently doing some meta research in his field. By crawling a publication database by specific keywords he got list of papers which need to be analyzed. However, most of the paper are hidden behind pay walls. Luckily there’s scihub. The most hated and beloved platform to get your hands on almost any scientific paper. He asked me wether I could help him building script that downloads papers from scihub based on a list of dois.

Sure I said, opended the browser and had a look at the requests that need to be performed to download a single paper.

When submitting the doi the following request is sent.

image

The http code 302 means that there is a redirection happening. If we want automate this process following redirects must be enabled.

image

The redirected response contains the paper. The url to the pdf file can be extracted using regex.

The initial request can be copied as curl command. This makes writing a script and bypassing any agent-checks much easier.

image

We both started Visual Studio Code and connected via the live coding plugin. My brother sent me a excerpt of the doi list:

doi-list.txt

10.3390/ijms21062239
10.1007/s12094-019-02276-8
10.31782/IJCRR.2019.11242
10.1016/j.chom.2019.05.007
10.3390/v11070638
10.1128/IAI.00733-18
10.1186/s12861-019-0183-y
10.1155/2019/8472712
10.1002/stem.2852
10.3390/genes9040176
10.1016/j.neulet.2018.01.040
10.1093/molehr/gax070
10.15252/emmm.201708213
10.1007/s40139-017-0137-7
10.1111/cas.13155
10.4103/0366-6999.191782
10.1126/science.aaf5211

And together we built this bash simple script:

scihub-download.sh

urlencode() {
    # urlencode <string>
    old_lc_collate=$LC_COLLATE
    LC_COLLATE=C

    local length="${#1}"
    for (( i = 0; i < length; i++ )); do
        local c="${1:i:1}"
        case $c in
            [a-zA-Z0-9.~_-]) printf "$c" ;;
            *) printf '%%%02X' "'$c" ;;
        esac
    done

    LC_COLLATE=$old_lc_collate
}

readarray -t list <doi-list.txt

for doi in "${list[@]}"
do
    echo "Download for doi: $(urlencode $doi)"
    link=$(curl -s -L 'https://sci-hub.tw/' --compressed \
        -H 'User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0' \
        -H 'Accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' \
        -H 'Accept-Language: en-US,en;q=0.5' \
        -H 'Content-Type: application/x-www-form-urlencoded' \
        -H 'Origin: https://sci-hub.tw' \
        -H 'DNT: 1' \
        -H 'Connection: keep-alive' \
        -H 'Referer: https://sci-hub.tw/' \
        -H 'Cookie: __ddg1=SFEVzNPdQpdmIWBwzsBq; session=45c4aaad919298b2eb754b6dd84ceb2d; refresh=1588795770.5886; __ddg2=6iYsE2844PoxLmj7' \
        -H 'Upgrade-Insecure-Requests: 1' \
        -H 'Pragma: no-cache' \
        -H 'Cache-Control: no-cache' \
        -H 'TE: Trailers' \
        --data "sci-hub-plugin-check=&request=$(urlencode $doi)". | grep -oP  "(?<=//).+(?=#)")

    echo "Found link: $link"
    curl -s -L $link --output $(urlencode $doi).pdf
done

It reads the text file and sends a request to scihub for each entry. If the response contains a valid pdf link, the file is downloaded. The doi request parameter must be url encoded.

In the next step he will process the content of the papers and analyze them for specific

©著作权归作者所有,转载或内容合作请联系作者
【社区内容提示】社区部分内容疑似由AI辅助生成,浏览时请结合常识与多方信息审慎甄别。
平台声明:文章内容(如有图片或视频亦包括在内)由作者上传并发布,文章内容仅代表作者本人观点,简书系信息发布平台,仅提供信息存储服务。