File size: 3,928 Bytes
1d4789e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
import argparse
import asyncio
import aiohttp

from tqdm import tqdm as bar

from bs4 import BeautifulSoup
from concurrent.futures import ProcessPoolExecutor
import re
from datasets import Dataset, DatasetDict
import json


parser = argparse.ArgumentParser(description='Process some arguments.')
parser.add_argument('--artist_id', type=int, help='Artist ID in Genius API')
parser.add_argument('--token', type=str, help='Genius API token')
parser.add_argument('--save_path', type=str, help='Path where to save the resulting json file')
args = parser.parse_args()


async def get_song_urls(artist_id):
    access_token = 'Bearer ' + args.token
    authorization_header = {'authorization': access_token}
    urls = []
    async with aiohttp.ClientSession(headers=authorization_header) as session:
        with bar(total=None) as pbar:
            pbar.set_description("⏳ Searching songs...")
            next_page = 1
            while next_page is not None:
                async with session.get(
                        f"https://api.genius.com/artists/{artist_id}/songs?sort=popularity&per_page=50&page={next_page}",
                        timeout=999) as resp:
                    response = await resp.json()
                    response = response['response']
                next_page = response['next_page']

                for song in response['songs']:
                    urls.append(song['url'])
                pbar.update(len(response['songs']))
            # pbar.set_description("✅ Done")
    return urls


def process_page(html):
    '''Meant for CPU-bound workload'''
    html = BeautifulSoup(html.replace('<br/>', '\n'), 'html.parser')
    div = html.find("div", class_=re.compile("^lyrics$|Lyrics__Root"))
    if div is None:
        lyrics = ""
    else:
        lyrics = div.get_text()

    lyrics = re.sub(r'(\[.*?\])*', '', lyrics)
    lyrics = re.sub('\n{2}', '\n', lyrics)  # Gaps between verses

    lyrics = str(lyrics.strip("\n"))
    lyrics = lyrics.replace("EmbedShare URLCopyEmbedCopy", "").replace("'", "")
    lyrics = re.sub("[\(\[].*?[\)\]]", "", lyrics)
    lyrics = re.sub(r'\d+$', '', lyrics)
    lyrics = str(lyrics).lstrip().rstrip()
    lyrics = str(lyrics).replace("\n\n", "\n")
    lyrics = str(lyrics).replace("\n\n", "\n")
    lyrics = re.sub(' +', ' ', lyrics)
    lyrics = str(lyrics).replace('"', "")
    # lyrics = str(lyrics).replace("'", "")
    lyrics = str(lyrics).replace("*", "")
    return lyrics


async def fetch_page(url, session):
    '''Meant for IO-bound workload'''
    async with session.get(url, timeout=999) as res:
        return await res.text()


async def process(url, session, pool, pbar):
    html = await fetch_page(url, session)
    pbar.update(1)
    return await asyncio.wrap_future(pool.submit(process_page, html))


async def parse_lyrics(urls):
    pool = ProcessPoolExecutor()
    pbar = bar(total=len(urls))
    async with aiohttp.ClientSession() as session:
        pbar.set_description("⏳ Parsing lyrics...")
        coros = (process(url, session, pool, pbar) for url in urls)
        return await asyncio.gather(*coros)


def create_dataset(lyrics):
  dataset = {}
  dataset['train'] = Dataset.from_dict({'text': list(lyrics)})
  datasets = DatasetDict(dataset)
  del dataset
  return datasets


def save_dataset(datasets):
    train = datasets['train'].to_dict()
    dataset_dict = {'train': train['text']}
    with open(args.save_path, 'w', encoding='utf-8') as fp:
        json.dump(dataset_dict, fp, ensure_ascii=False)


def main():

    loop = asyncio.get_event_loop()
    urls = loop.run_until_complete(get_song_urls(args.artist_id))
    lyrics = loop.run_until_complete(parse_lyrics(urls))
    loop.close()

    dataset = create_dataset(lyrics)
    save_dataset(dataset)


if __name__ == '__main__':
    main()