Script to backup Last.fm scrobbles
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 

138 lines
3.8 KiB

#!/usr/bin/env python3
import json
import csv
import os.path
import requests
__author__ = "Alexander Popov"
__version__ = "2.1.0"
__license__ = "Unlicense"
def get_pages(username, api_key, limit=200):
""" Getting the number of pages with scrobbling data """
response = requests.get(
"https://ws.audioscrobbler.com/2.0/"
"?method=user.getrecenttracks&user="
"{0}&api_key={1}&format=json&limit={2}".format(username, api_key, limit)
)
data = json.loads(response.content.decode("utf8"))
return int(data["recenttracks"]["@attr"]["totalPages"])
def get_page(username, api_key, page, limit=200):
""" Getting scrobbling data from a page """
response = requests.get(
"https://ws.audioscrobbler.com/2.0/"
"?method=user.getrecenttracks&user={0}&api_key={1}&format=json"
"&limit={2}&page={3}".format(username, api_key, limit, page)
)
data = json.loads(response.content.decode("utf8"))
return data["recenttracks"]["track"]
def get_now_scrobbling(username, api_key):
""" Getting now scrobbling track """
response = requests.get(
"https://ws.audioscrobbler.com/2.0/"
"?method=user.getrecenttracks&user={0}"
"&api_key={1}&format=json&limit=1".format(username, api_key)
)
data = json.loads(response.content.decode("utf-8"))
if "@attr" in data["recenttracks"]["track"][0]:
return True
else:
return False
def scrobbling_export(tracks, username, export_format="dump"):
""" Save scrobbled track via various format """
if export_format == "dump":
with open("%s.json" % (username), "w", encoding="utf-8") as f:
data = json.dumps(tracks, indent=4, sort_keys=True, ensure_ascii=False)
f.write(data)
elif export_format == "simple":
_ = []
for track in tracks:
_.append(
{
"artist": track["artist"]["#text"],
"name": track["name"],
"album": track["album"]["#text"],
"date": int(track["date"]["uts"]),
}
)
with open("%s.json" % (username), "w", encoding="utf-8") as f:
data = json.dumps(_, indent=4, sort_keys=True, ensure_ascii=False)
f.write(data)
elif export_format == "csv":
_ = []
for track in tracks:
_.append(
[
track["artist"]["#text"],
track["name"],
track["album"]["#text"],
int(track["date"]["uts"]),
]
)
with open("%s.csv" % (username), "w", encoding="utf-8", newline="") as f:
data = csv.writer(f, quoting=csv.QUOTE_NONNUMERIC, delimiter=",")
data.writerow(["artist", "track", "album", "date"])
data.writerows(_)
return 1
if __name__ == "__main__":
_ = dict()
if os.path.exists("./config.json"):
with open("./config.json") as f:
_ = json.loads(f.read())
else:
_["api_key"] = input("API Key: ")
_["username"] = input("Username: ")
api_key, username = _["api_key"], _["username"]
total_pages = get_pages(username, api_key)
current_page = 1
scrobbled = []
while current_page <= total_pages:
print(
"\r{0:.2f}% [{1} of {2}]".format(
(current_page * 100 / total_pages), current_page, total_pages
),
end="",
)
response = get_page(username, api_key, current_page)
for track in response:
scrobbled.append(track)
current_page += 1
# if get_now_scrobbling(username, api_key):
# scrobbled.pop(0)
if scrobbling_export(scrobbled, username, _["export_format"]):
print("\n{0} tracks saved!".format(len(scrobbled), username))