File size: 1,772 Bytes
3cfadc8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
import csv
import numpy as np
import sys
import pandas as pd
import re
sys.path.append("./")
from evaluations.utils import *

token = os.getenv("githubToken")
custom_order = ["MICCAI", "MIDL", "Nature", "arXiv"]

for venue in custom_order:
    df = pd.read_excel("https://docs.google.com/spreadsheets/d/e/2PACX-1vQjpsSYcEcYUVB-88bCQ01UfQf0z9m16ax7p1ft03G68Nr-DdXHpPt-xOFSrXFj1N49AjK5nYhmKBfo/pub?output=xlsx", sheet_name=venue)
    df = df.replace('\t', ' ', regex=True)
    df.to_csv(f'data/{venue}.csv', sep="\t")

# Store all evaluations here
paper_dump = pd.DataFrame()
# Official color codes for conferences
zip_idx = 0

for venue in custom_order:
    with open(f'data/{venue}.csv') as file:
        tsv_file = csv.reader(file, delimiter="\t")
        for row in tsv_file:
            if (row[0] == ""):
                continue

            if (row[1] == ""):
                continue

            repo_url = row[4]
            username, repo_name = decompose_url(repo_url)
            repo_save_name = f"repo_{zip_idx}.zip"
            repository_zip_name = f"data/test/{repo_save_name}"
            log(0, "LOG", f"Fetching github repository: https://github.com/{username}/{repo_name}")
            fetch_repo(0, repo_url, repository_zip_name, token)

            if (os.path.exists(repository_zip_name)):
                paper_dump = pd.concat([paper_dump, pd.DataFrame({"venue": venue, "title": [row[1]], "year": [row[2]], "pdf": [row[3]], "url": [row[4]], "public": [row[5]], "dependencies": [row[6]], "training": [row[7]], "evaluation": [row[8]], "weights": [row[9]], "readme": [row[10]], "license": [row[11]], "zip_idx": [ repository_zip_name ]})], ignore_index=True)
                zip_idx += 1
            
paper_dump.to_csv(f'data/zipfiles.csv', sep="\t")