# vim: set fileencoding=utf-8 import csv import argparse import copy import io import jaconv # type: ignore import json import os import struct import xml.etree.ElementTree as ET from pathlib import Path from sqlalchemy.engine import CursorResult # type: ignore from sqlalchemy.orm import sessionmaker # type: ignore from sqlalchemy.sql import text # type: ignore from sqlalchemy.exc import IntegrityError # type: ignore from typing import Any, Callable, Dict, List, Optional, Tuple from bemani.common import ( GameConstants, VersionConstants, DBConstants, PEFile, InvalidOffsetException, Time, ) from bemani.format import ARC, IFS, IIDXChart, IIDXMusicDB from bemani.data import Config, Server, Song from bemani.data.interfaces import APIProviderInterface from bemani.data.api.music import GlobalMusicData from bemani.data.api.game import GlobalGameData from bemani.data.mysql.music import MusicData from bemani.data.mysql.user import UserData from bemani.utils.config import load_config class CLIException(Exception): pass class ReadAPI(APIProviderInterface): def __init__(self, server: str, token: str) -> None: self.__server = server self.__token = token def get_all_servers(self) -> List[Server]: return [ Server( 0, Time.now(), self.__server, self.__token, False, False, ) ] class ImportBase: def __init__( self, config: Config, game: GameConstants, version: Optional[int], no_combine: bool, update: bool, ) -> None: self.game = game self.version = version self.update = update self.no_combine = no_combine self.__config = config self.__engine = self.__config.database.engine self.__sessionmanager = sessionmaker(self.__engine) self.__conn = self.__engine.connect() self.__session = self.__sessionmanager(bind=self.__conn) self.__batch = False def start_batch(self) -> None: self.__batch = True def finish_batch(self) -> None: self.__session.commit() self.__batch = False def execute( self, sql: str, params: Optional[Dict[str, Any]] = None ) -> CursorResult: if not self.__batch: raise Exception("Logic error, cannot execute outside of a batch!") if self.__config.database.read_only: # See if this is an insert/update/delete for write_statement in [ "insert into ", "update ", "delete from ", ]: if write_statement in sql.lower(): raise Exception("Read-only mode is active!") return self.__session.execute(text(sql), params if params is not None else {}) def remote_music(self, server: str, token: str) -> GlobalMusicData: api = ReadAPI(server, token) user = UserData(self.__config, self.__session) music = MusicData(self.__config, self.__session) return GlobalMusicData(api, user, music) def remote_game(self, server: str, token: str) -> GlobalGameData: api = ReadAPI(server, token) return GlobalGameData(api) def get_next_music_id(self) -> int: cursor = self.execute("SELECT MAX(id) AS next_id FROM `music`") result = cursor.fetchone() try: return result["next_id"] + 1 except TypeError: # Nothing in DB return 1 def get_music_id_for_song( self, songid: int, chart: int, version: Optional[int] = None ) -> Optional[int]: if version is None: # Normal lookup if self.version is None: raise Exception( "Cannot get music ID for song when operating on all versions!" ) version = self.version sql = "SELECT id FROM `music` WHERE songid = :songid AND chart = :chart AND game = :game AND version != :version" else: # Specific version lookup sql = "SELECT id FROM `music` WHERE songid = :songid AND chart = :chart AND game = :game AND version = :version" cursor = self.execute( sql, { "songid": songid, "chart": chart, "game": self.game.value, "version": version, }, ) if cursor.rowcount != 0: result = cursor.fetchone() return result["id"] else: return None def get_music_id_for_song_data( self, title: Optional[str], artist: Optional[str], genre: Optional[str], chart: int, version: Optional[int] = None, ) -> Optional[int]: frags = [] if title is not None: frags.append("name = :title") if artist is not None: frags.append("artist = :artist") if genre is not None: frags.append("genre = :genre") frags.append("chart = :chart") frags.append("game = :game") if version is None: # Normal lookup if self.version is None: raise Exception( "Cannot get music ID for song when operating on all versions!" ) version = self.version frags.append("version != :version") else: frags.append("version = :version") sql = "SELECT id FROM `music` WHERE " + " AND ".join(frags) cursor = self.execute( sql, { "title": title, "artist": artist, "genre": genre, "chart": chart, "game": self.game.value, "version": version, }, ) if cursor.rowcount != 0: result = cursor.fetchone() return result["id"] else: return None def insert_music_id_for_song( self, musicid: int, songid: int, chart: int, name: Optional[str] = None, artist: Optional[str] = None, genre: Optional[str] = None, data: Optional[Dict[str, Any]] = None, version: Optional[int] = None, ) -> None: version = version if version is not None else self.version if version is None: raise Exception( "Cannot get insert new song when operating on all versions!" ) if data is None: jsondata = "{}" else: jsondata = json.dumps(data) try: sql = ( "INSERT INTO `music` (id, songid, chart, game, version, name, artist, genre, data) " + "VALUES (:id, :songid, :chart, :game, :version, :name, :artist, :genre, :data)" ) self.execute( sql, { "id": musicid, "songid": songid, "chart": chart, "game": self.game.value, "version": version, "name": name, "artist": artist, "genre": genre, "data": jsondata, }, ) except IntegrityError: if self.update: print("Entry already existed, so updating information!") self.update_metadata_for_song( songid, chart, name, artist, genre, data, version ) else: print("Entry already existed, so skip creating a second one!") def update_metadata_for_song( self, songid: int, chart: int, name: Optional[str] = None, artist: Optional[str] = None, genre: Optional[str] = None, data: Optional[Dict[str, Any]] = None, version: Optional[int] = None, ) -> None: if data is None: jsondata = None else: jsondata = json.dumps(data) version = version if version is not None else self.version updates = [] if jsondata is not None: updates.append("data = :data") if name is not None: updates.append("name = :name") if artist is not None: updates.append("artist = :artist") if genre is not None: updates.append("genre = :genre") if len(updates) == 0: return sql = f"UPDATE `music` SET {', '.join(updates)} WHERE songid = :songid AND chart = :chart AND game = :game" if version is not None: sql = sql + " AND version = :version" self.execute( sql, { "songid": songid, "chart": chart, "game": self.game.value, "version": version, "name": name, "artist": artist, "genre": genre, "data": jsondata, }, ) def update_metadata_for_music_id( self, musicid: int, name: Optional[str] = None, artist: Optional[str] = None, genre: Optional[str] = None, data: Optional[Dict[str, Any]] = None, version: Optional[int] = None, ) -> None: if data is None: jsondata = None else: jsondata = json.dumps(data) version = version if version is not None else self.version updates = [] if jsondata is not None: updates.append("data = :data") if name is not None: updates.append("name = :name") if artist is not None: updates.append("artist = :artist") if genre is not None: updates.append("genre = :genre") if len(updates) == 0: return sql = f"UPDATE `music` SET {', '.join(updates)} WHERE id = :musicid AND game = :game" if version is not None: sql = sql + " AND version = :version" self.execute( sql, { "musicid": musicid, "game": self.game.value, "version": version, "name": name, "artist": artist, "genre": genre, "data": jsondata, }, ) def insert_catalog_entry( self, cattype: str, catid: int, data: Optional[Dict[str, Any]] = None, ) -> None: if data is None: jsondata = "{}" else: jsondata = json.dumps(data) try: sql = ( "INSERT INTO `catalog` (game, version, type, id, data) " + "VALUES (:game, :version, :type, :id, :data)" ) self.execute( sql, { "id": catid, "type": cattype, "game": self.game.value, "version": self.version, "data": jsondata, }, ) except IntegrityError: if self.update: print("Entry already existed, so updating information!") sql = ( "UPDATE `catalog` SET data = :data WHERE " + "game = :game AND version = :version AND type = :type AND id = :id" ) self.execute( sql, { "id": catid, "type": cattype, "game": self.game.value, "version": self.version, "data": jsondata, }, ) else: print("Entry already existed, so skip creating a second one!") def close(self) -> None: """ Close any open data connection. """ # Make sure we don't leak connections after finising insertion. if self.__batch: raise Exception("Logic error, opened a batch without closing!") if self.__session is not None: self.__session.close() if self.__conn is not None: self.__conn.close() self.__conn = None if self.__engine is not None: self.__engine.dispose() self.__engine = None class PopnScrapeConfiguration: def __init__( self, *, version: str, offset: int, step: int, length: int, file_offset: int, file_step: int, genre_offset: int, title_offset: int, artist_offset: int, comment_offset: int, english_title_offset: Optional[int], english_artist_offset: Optional[int], extended_genre_offset: Optional[int], charts_offset: int, folder_offset: int, easy_offset: int, normal_offset: int, hyper_offset: int, ex_offset: int, battle_normal_offset: int, battle_hyper_offset: int, easy_file_offset: int, normal_file_offset: int, hyper_file_offset: int, ex_file_offset: int, battle_normal_file_offset: int, battle_hyper_file_offset: int, packedfmt: str, file_folder_offset: int, file_name_offset: int, filefmt: str, available_charts: Callable[[int], Tuple[bool, bool, bool, bool, bool, bool]], ) -> None: self.version = version self.offset = offset self.step = step self.length = length self.file_offset = file_offset self.file_step = file_step self.genre_offset = genre_offset self.title_offset = title_offset self.artist_offset = artist_offset self.comment_offset = comment_offset self.english_title_offset = english_title_offset self.english_artist_offset = english_artist_offset self.extended_genre_offset = extended_genre_offset self.charts_offset = charts_offset self.folder_offset = folder_offset self.easy_offset = easy_offset self.normal_offset = normal_offset self.hyper_offset = hyper_offset self.ex_offset = ex_offset self.battle_normal_offset = battle_normal_offset self.battle_hyper_offset = battle_hyper_offset self.easy_file_offset = easy_file_offset self.normal_file_offset = normal_file_offset self.hyper_file_offset = hyper_file_offset self.ex_file_offset = ex_file_offset self.battle_normal_file_offset = battle_normal_file_offset self.battle_hyper_file_offset = battle_hyper_file_offset self.packedfmt = packedfmt self.file_folder_offset = file_folder_offset self.file_name_offset = file_name_offset self.filefmt = filefmt self.available_charts = available_charts class ImportPopn(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: actual_version = { "19": VersionConstants.POPN_MUSIC_TUNE_STREET, "20": VersionConstants.POPN_MUSIC_FANTASIA, "21": VersionConstants.POPN_MUSIC_SUNNY_PARK, "22": VersionConstants.POPN_MUSIC_LAPISTORIA, "23": VersionConstants.POPN_MUSIC_ECLALE, "24": VersionConstants.POPN_MUSIC_USANEKO, "25": VersionConstants.POPN_MUSIC_PEACE, "26": VersionConstants.POPN_MUSIC_KAIMEI_RIDDLES, "omni-24": VersionConstants.POPN_MUSIC_USANEKO + DBConstants.OMNIMIX_VERSION_BUMP, # Omnimix v2 only works for 24 - 26 "omni-25": VersionConstants.POPN_MUSIC_PEACE + DBConstants.OMNIMIX_VERSION_BUMP, "omni-26": VersionConstants.POPN_MUSIC_KAIMEI_RIDDLES + DBConstants.OMNIMIX_VERSION_BUMP, }.get(version, -1) if actual_version == VersionConstants.POPN_MUSIC_TUNE_STREET: # Pop'n 19 has extra charts for old play modes (challenge and enjoy mode). # Cho challenge is analogous to regular mode in newer games, but Pop'n # 19 doesn't have easy charts, just 5 button charts. self.charts = [1, 2, 3, 4, 5, 6, 7, 8, 9] elif actual_version >= VersionConstants.POPN_MUSIC_FANTASIA: # Newer pop'n has charts for easy, normal, hyper, another self.charts = [0, 1, 2, 3] else: raise CLIException( "Unsupported Pop'n Music version, expected one of the following: 19, 20, 21, 22, 23, 24, omni-24, 25, omni-25, 26, omni-26!" ) super().__init__( config, GameConstants.POPN_MUSIC, actual_version, no_combine, update ) def scrape_xml( self, xmlfile: Path, songs: List[Dict[str, Any]] = [] ) -> List[Dict[str, Any]]: with open(xmlfile, "rb") as xmlhandle: xmldata = xmlhandle.read().decode("shift_jisx0213") root = ET.fromstring(xmldata) for music_entry in root.findall("music"): difficulties = [0, 0, 0, 0, 0, 0] filenames = ["", "", "", "", "", ""] diff_map = { "ep": 0, "np": 1, "hp": 2, "op": 3, "bp_n": 4, "bp_h": 5, } charts = music_entry.find("charts") if charts is not None: for chart in charts.findall("chart"): chart_idx = diff_map.get(chart.attrib["idx"]) if chart.find("diff") is not None: difficulties[chart_idx] = int(chart.find("diff").text) filenames[ chart_idx ] = f'{chart.find("folder").text}/{chart.find("filename").text}' songinfo: Dict # Check if song metadata is in this entry if music_entry.find("fw_title") is not None: songinfo = { "id": int(music_entry.attrib["id"]), "title": music_entry.find("fw_title").text, "artist": music_entry.find("fw_artist").text, "genre": music_entry.find("fw_genre").text, "comment": music_entry.find("genre").text, "title_en": music_entry.find("title").text, "artist_en": music_entry.find("artist").text, "long_genre": "", "folder": music_entry.find("folder").text, "difficulty": { "standard": { "easy": difficulties[0], "normal": difficulties[1], "hyper": difficulties[2], "ex": difficulties[3], }, "battle": { "normal": difficulties[4], "hyper": difficulties[5], }, }, "file": { "standard": { "easy": filenames[0], "normal": filenames[1], "hyper": filenames[2], "ex": filenames[3], }, "battle": { "normal": filenames[4], "hyper": filenames[5], }, }, } # It's not here so find the entry at the current song id else: for song in songs: if song["id"] == int(music_entry.attrib["id"]): if difficulties is not None: for diff, i in zip( ["easy", "normal", "hyper", "ex"], range(4) ): song["difficulty"]["standard"][diff] = ( difficulties[i] if difficulties[i] else song["difficulty"]["standard"][diff] ) song["file"]["standard"][diff] = ( filenames[i] if filenames[i] else song["file"]["standard"][diff] ) song["difficulty"]["battle"]["normal"] = ( difficulties[4] if difficulties[4] else song["difficulty"]["battle"]["normal"] ) song["difficulty"]["battle"]["hyper"] = ( difficulties[5] if difficulties[5] else song["difficulty"]["battle"]["hyper"] ) song["file"]["battle"]["normal"] = ( filenames[4] if filenames[4] else song["file"]["battle"]["normal"] ) song["file"]["battle"]["hyper"] = ( filenames[5] if filenames[5] else song["file"]["battle"]["hyper"] ) else: song["genre"] = music_entry.find("fw_genre").text song["comment"] = music_entry.find("genre").text break continue # Fix accent issues with title/artist accent_lut: Dict[str, str] = { "鵝": "7", "圄": "à", "圉": "ä", "鵤": "Ä", "鵑": "👁", "鶤": "©", "圈": "é", "鵐": "ê", "鵙": "Ə", "鵲": "ë", "!": "!", "囿": "♥", "鶚": "㊙", "鶉": "ó", "鶇": "ö", "鶲": "Ⓟ", "鶫": "²", "圍": "@", "圖": "ţ", "鵺": "Ü", "囎": ":", "囂": "♡", "釁": "🐾", } for orig, rep in accent_lut.items(): songinfo["title"] = songinfo["title"].replace(orig, rep) songinfo["artist"] = songinfo["artist"].replace(orig, rep) songinfo["title_en"] = songinfo["title_en"].replace(orig, rep) songinfo["artist_en"] = songinfo["artist_en"].replace(orig, rep) songinfo["genre"] = songinfo["genre"].replace(orig, rep) songs.append(songinfo) return songs def scrape(self, infile: str) -> List[Dict[str, Any]]: with open(infile, mode="rb") as myfile: data = myfile.read() myfile.close() pe = PEFile(data) configurations: List[PopnScrapeConfiguration] = [] if self.version == VersionConstants.POPN_MUSIC_TUNE_STREET: # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( True, # Always an easy chart True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on K39:J:A:A:2011061400 configurations.append( PopnScrapeConfiguration( version="K39:J:A:A:2011061400", # Normal offset for music DB, size offset=0x1F68E8, step=72, length=1048, # Offset and step of file DB file_offset=0x2D6888, file_step=24, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=None, english_artist_offset=None, extended_genre_offset=None, charts_offset=6, folder_offset=7, # Offsets for normal chart difficulties easy_offset=12, normal_offset=10, hyper_offset=11, ex_offset=13, # Offsets for battle chart difficulties battle_normal_offset=14, battle_hyper_offset=15, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=16, hyper_file_offset=17, ex_file_offset=19, battle_normal_file_offset=20, battle_hyper_file_offset=21, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event flags? "B" # Event flags? "B" # Normal difficulty "B" # Hyper difficulty "B" # Easy difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "x" # ?? "x" # ?? "x" # ?? "H" # Normal chart pointer "H" # Hyper chart pointer "H" # Easy chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_FANTASIA: # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( True, # Always an easy chart True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on L39:J:A:A:2011121400 configurations.append( PopnScrapeConfiguration( version="L39:J:A:A:2011121400", # Normal offset for music DB, size offset=0x1797D0, step=160, length=1116, # Offset and step of file DB file_offset=0x238AD0, file_step=24, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=None, english_artist_offset=None, extended_genre_offset=None, charts_offset=6, folder_offset=7, # Offsets for normal chart difficulties easy_offset=12, normal_offset=10, hyper_offset=11, ex_offset=13, # Offsets for battle chart difficulties battle_normal_offset=14, battle_hyper_offset=15, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=16, hyper_file_offset=17, ex_file_offset=19, battle_normal_file_offset=20, battle_hyper_file_offset=21, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event flags? "B" # Event flags? "B" # Normal difficulty "B" # Hyper difficulty "B" # Easy difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "x" # ?? "x" # ?? "x" # ?? "H" # Normal chart pointer "H" # Hyper chart pointer "H" # Easy chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on L39:J:A:A:2012091900 configurations.append( PopnScrapeConfiguration( version="L39:J:A:A:2012091900", # Normal offset for music DB, size offset=0x1AE240, step=160, length=1122, # Offset and step of file DB file_offset=0x273768, file_step=24, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=None, english_artist_offset=None, extended_genre_offset=None, charts_offset=6, folder_offset=7, # Offsets for normal chart difficulties easy_offset=12, normal_offset=10, hyper_offset=11, ex_offset=13, # Offsets for battle chart difficulties battle_normal_offset=14, battle_hyper_offset=15, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=16, hyper_file_offset=17, ex_file_offset=19, battle_normal_file_offset=20, battle_hyper_file_offset=21, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event flags? "B" # Event flags? "B" # Normal difficulty "B" # Hyper difficulty "B" # Easy difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "x" # ?? "x" # ?? "x" # ?? "H" # Normal chart pointer "H" # Hyper chart pointer "H" # Easy chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_SUNNY_PARK: # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2012120900 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2012120900", # Normal offset for music DB, size offset=0x16C880, step=164, length=1184, # Offset and step of file DB file_offset=0x2399B8, file_step=28, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=6, charts_offset=9, folder_offset=10, # Offsets for normal chart difficulties easy_offset=13, normal_offset=14, hyper_offset=15, ex_offset=16, # Offsets for battle chart difficulties battle_normal_offset=17, battle_hyper_offset=18, # Offsets into which offset to seek to for file lookups easy_file_offset=19, normal_file_offset=20, hyper_file_offset=21, ex_file_offset=22, battle_normal_file_offset=21, battle_hyper_file_offset=22, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "I" # Extended genre? "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2012121200 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2012121200", # Normal offset for music DB, size offset=0x16C880, step=164, length=1183, # Offset and step of file DB file_offset=0x2399B8, file_step=28, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=6, charts_offset=9, folder_offset=10, # Offsets for normal chart difficulties easy_offset=13, normal_offset=14, hyper_offset=15, ex_offset=16, # Offsets for battle chart difficulties battle_normal_offset=17, battle_hyper_offset=18, # Offsets into which offset to seek to for file lookups easy_file_offset=19, normal_file_offset=20, hyper_file_offset=21, ex_file_offset=22, battle_normal_file_offset=21, battle_hyper_file_offset=22, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "I" # Extended genre? "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2013011500 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2013011500", # Normal offset for music DB, size offset=0x170ED8, step=164, length=1183, # Offset and step of file DB file_offset=0x23EFC0, file_step=28, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=6, charts_offset=9, folder_offset=10, # Offsets for normal chart difficulties easy_offset=13, normal_offset=14, hyper_offset=15, ex_offset=16, # Offsets for battle chart difficulties battle_normal_offset=17, battle_hyper_offset=18, # Offsets into which offset to seek to for file lookups easy_file_offset=19, normal_file_offset=20, hyper_file_offset=21, ex_file_offset=22, battle_normal_file_offset=21, battle_hyper_file_offset=22, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "I" # Extended genre? "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2014061900 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2014061900", # Normal offset for music DB, size offset=0x1FB640, step=164, length=1280, # Offset and step of file DB file_offset=0x2E0D20, file_step=28, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=6, charts_offset=9, folder_offset=10, # Offsets for normal chart difficulties easy_offset=13, normal_offset=14, hyper_offset=15, ex_offset=16, # Offsets for battle chart difficulties battle_normal_offset=17, battle_hyper_offset=18, # Offsets into which offset to seek to for file lookups easy_file_offset=19, normal_file_offset=20, hyper_file_offset=21, ex_file_offset=22, battle_normal_file_offset=21, battle_hyper_file_offset=22, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "I" # Extended genre? "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_LAPISTORIA: # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2015081900 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2015081900", # Normal offset for music DB, size offset=0x3124B0, step=160, length=1423, # Offset and step of file DB file_offset=0x472130, file_step=28, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_ECLALE: # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2016100500 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2016100500", # Normal offset for music DB, size offset=0x2DE5C8, step=160, length=1551, # Offset and step of file DB file_offset=0x2D1948, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "H" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_USANEKO or self.version == ( VersionConstants.POPN_MUSIC_USANEKO + DBConstants.OMNIMIX_VERSION_BUMP ): # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2018082100 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2018082100", # Normal offset for music DB, size offset=0x299410, step=172, length=1704, # Offset and step of file DB file_offset=0x28B108, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2018101500 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2018101500", # Normal offset for music DB, size offset=0x299210, step=172, length=1704, # Offset and step of file DB file_offset=0x28AF08, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif self.version == VersionConstants.POPN_MUSIC_PEACE or self.version == ( VersionConstants.POPN_MUSIC_PEACE + DBConstants.OMNIMIX_VERSION_BUMP ): # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2019042300 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2019042300", # Normal offset for music DB, size offset=0x2B3840, step=172, length=1780, # Offset and step of file DB file_offset=0x2A48F8, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2020092800 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2020092800", # Normal offset for music DB, size offset=0x2C7C78, step=172, length=1877, # Offset and step of file DB file_offset=0x2B8010, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) elif ( self.version == VersionConstants.POPN_MUSIC_KAIMEI_RIDDLES or self.version == ( VersionConstants.POPN_MUSIC_KAIMEI_RIDDLES + DBConstants.OMNIMIX_VERSION_BUMP ) ): # Decoding function for chart masks def available_charts( mask: int, ) -> Tuple[bool, bool, bool, bool, bool, bool]: return ( mask & 0x0080000 > 0, # Easy chart bit True, # Always a normal chart mask & 0x1000000 > 0, # Hyper chart bit mask & 0x2000000 > 0, # Ex chart bit True, # Always a battle normal chart mask & 0x4000000 > 0, # Battle hyper chart bit ) # Based on M39:J:A:A:2021042600 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2021042600", # Normal offset for music DB, size offset=0x2D0628, step=172, length=1945, # Offset and step of file DB file_offset=0x2C00C0, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2022042500 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2022042500", # Normal offset for music DB, size offset=0x2DB398, step=172, length=2012, # Offset and step of file DB file_offset=0x2CA510, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) # Based on M39:J:A:A:2022061300 configurations.append( PopnScrapeConfiguration( version="M39:J:A:A:2022061300", # Normal offset for music DB, size offset=0x2DEA68, step=172, length=2019, # Offset and step of file DB file_offset=0x2CDB00, file_step=32, # Standard lookups genre_offset=0, title_offset=1, artist_offset=2, comment_offset=3, english_title_offset=4, english_artist_offset=5, extended_genre_offset=None, charts_offset=8, folder_offset=9, # Offsets for normal chart difficulties easy_offset=12, normal_offset=13, hyper_offset=14, ex_offset=15, # Offsets for battle chart difficulties battle_normal_offset=16, battle_hyper_offset=17, # Offsets into which offset to seek to for file lookups easy_file_offset=18, normal_file_offset=19, hyper_file_offset=20, ex_file_offset=21, battle_normal_file_offset=22, battle_hyper_file_offset=23, packedfmt=( "<" "I" # Genre "I" # Title "I" # Artist "I" # Comment "I" # English Title "I" # English Artist "H" # ?? "H" # ?? "I" # Available charts mask "I" # Folder "I" # Event unlocks? "I" # Event unlocks? "B" # Easy difficulty "B" # Normal difficulty "B" # Hyper difficulty "B" # EX difficulty "B" # Battle normal difficulty "B" # Battle hyper difficulty "xx" # Unknown pointer "H" # Easy chart pointer "H" # Normal chart pointer "H" # Hyper chart pointer "H" # EX chart pointer "H" # Battle normal pointer "H" # Battle hyper pointer "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx" ), # Offsets into file DB for finding file and folder. file_folder_offset=0, file_name_offset=1, filefmt="<" "I" "I" "I" "I" "I" "I" "I" "I", # Folder # Filename available_charts=available_charts, ) ) else: raise CLIException(f"Unsupported version {self.version}") def read_string(offset: int) -> str: # First, translate load offset in memory to disk offset offset = pe.virtual_to_physical(offset) # Now, grab bytes until we're null-terminated bytestring = [] while data[offset] != 0: bytestring.append(data[offset]) offset = offset + 1 # Its shift-jis encoded, so decode it now return bytes(bytestring).decode("shift_jisx0213") def file_chunk(config: PopnScrapeConfiguration, offset: int) -> Tuple[Any, ...]: fileoffset = config.file_offset + (config.file_step * offset) filedata = data[fileoffset : (fileoffset + config.file_step)] return struct.unpack(config.filefmt, filedata) def file_handle(config: PopnScrapeConfiguration, offset: int) -> str: chunk = file_chunk(config, offset) return ( read_string(chunk[config.file_folder_offset]) + "/" + read_string(chunk[config.file_name_offset]) ) for config in configurations: try: print(f"Trying configuration for game version {config.version}...") songs = [] for songid in range(config.length): chunkoffset = config.offset + (config.step * songid) chunkdata = data[chunkoffset : (chunkoffset + config.step)] unpacked = struct.unpack(config.packedfmt, chunkdata) valid_charts = config.available_charts( unpacked[config.charts_offset] ) songinfo = { "id": songid, "title": read_string(unpacked[config.title_offset]), "artist": read_string(unpacked[config.artist_offset]), "genre": read_string(unpacked[config.genre_offset]), "comment": read_string(unpacked[config.comment_offset]), "title_en": read_string(unpacked[config.english_title_offset]) if config.english_title_offset is not None else "", "artist_en": read_string(unpacked[config.english_artist_offset]) if config.english_artist_offset is not None else "", "long_genre": read_string( unpacked[config.extended_genre_offset] ) if config.extended_genre_offset is not None else "", "folder": unpacked[config.folder_offset], "difficulty": { "standard": { "easy": unpacked[config.easy_offset] if valid_charts[0] else 0, "normal": unpacked[config.normal_offset] if valid_charts[1] else 0, "hyper": unpacked[config.hyper_offset] if valid_charts[2] else 0, "ex": unpacked[config.ex_offset] if valid_charts[3] else 0, }, "battle": { "normal": unpacked[config.battle_normal_offset] if valid_charts[4] else 0, "hyper": unpacked[config.battle_hyper_offset] if valid_charts[5] else 0, }, }, "file": { "standard": { "easy": file_handle( config, unpacked[config.easy_file_offset] ) if valid_charts[0] else "", "normal": file_handle( config, unpacked[config.normal_file_offset] ) if valid_charts[1] else "", "hyper": file_handle( config, unpacked[config.hyper_file_offset] ) if valid_charts[2] else "", "ex": file_handle( config, unpacked[config.ex_file_offset] ) if valid_charts[3] else "", }, "battle": { "normal": file_handle( config, unpacked[config.battle_normal_file_offset] ) if valid_charts[4] else "", "hyper": file_handle( config, unpacked[config.battle_hyper_file_offset] ) if valid_charts[5] else "", }, }, } if ( songinfo["title"] in ["-", "‐"] and songinfo["genre"] in ["-", "‐"] and songinfo["artist"] in ["-", "‐"] and songinfo["comment"] in ["-", "‐"] ): # This is a removed song continue if ( songinfo["title"] == "DUMMY" and songinfo["artist"] == "DUMMY" and songinfo["genre"] == "DUMMY" ): # This is a song the intern left in continue # Fix accent issues with title/artist accent_lut: Dict[str, str] = { "鵝": "7", "圄": "à", "圉": "ä", "鵤": "Ä", "鵑": "👁", "鶤": "©", "圈": "é", "鵐": "ê", "鵙": "Ə", "鵲": "ë", "!": "!", "囿": "♥", "鶚": "㊙", "鶉": "ó", "鶇": "ö", "鶲": "Ⓟ", "鶫": "²", "圍": "@", "圖": "ţ", "鵺": "Ü", "囎": ":", "囂": "♡", "釁": "🐾", } for orig, rep in accent_lut.items(): songinfo["title"] = songinfo["title"].replace(orig, rep) songinfo["artist"] = songinfo["artist"].replace(orig, rep) songinfo["title_en"] = songinfo["title_en"].replace(orig, rep) songinfo["artist_en"] = songinfo["artist_en"].replace(orig, rep) songinfo["genre"] = songinfo["genre"].replace(orig, rep) songs.append(songinfo) # If we got here, that means we ran into no issues and didn't have to attempt another offset. print("Successfully parsed game DB!") return songs except (UnicodeError, InvalidOffsetException): # These offsets are possibly not correct, so try the next configuration. print("Failed to parse game DB using current inferred data version!") pass raise CLIException( f"Could not determine correct binary parser configuration for Pop'n Music version {self.version}" ) def lookup(self, server: str, token: str) -> List[Dict[str, Any]]: # Grab music info from remote server music = self.remote_music(server, token) songs = music.get_all_songs(self.game, self.version) lut: Dict[int, Dict[str, Any]] = {} chart_map = { 0: "easy", 1: "normal", 2: "hyper", 3: "ex", } # Format it the way we expect for song in songs: if song.chart not in chart_map: # Ignore charts on songs we don't support/care about. continue if song.id not in lut: lut[song.id] = { "id": song.id, "title": song.name, "artist": song.artist, "genre": song.genre, "comment": "", "title_en": "", "artist_en": "", "long_genre": "", "folder": song.data.get_str("category"), "difficulty": { "standard": { "easy": 0, "normal": 0, "hyper": 0, "ex": 0, }, "battle": { "normal": 0, "hyper": 0, }, }, "file": { "standard": { "easy": "", "normal": "", "hyper": "", "ex": "", }, "battle": { "normal": "", "hyper": "", }, }, } lut[song.id]["difficulty"]["standard"][ chart_map[song.chart] ] = song.data.get_int("difficulty") # Return the reassembled data return [val for _, val in lut.items()] def import_music_db(self, songs: List[Dict[str, Any]]) -> None: chart_map = { 0: "easy", 1: "normal", 2: "hyper", 3: "ex", } for song in songs: self.start_batch() for chart in self.charts: # First, try to find in the DB from another version old_id = self.get_music_id_for_song(song["id"], chart) # Now, look up metadata title = song["title_en"] if len(song["title_en"]) > 0 else song["title"] artist = ( song["artist_en"] if len(song["artist_en"]) > 0 else song["artist"] ) genre = song["genre"] # We only care about easy/normal/hyper/ex, so only provide mappings there if chart in chart_map: difficulty = song["difficulty"]["standard"][chart_map[chart]] else: difficulty = 0 if self.no_combine or old_id is None: # Insert original print( f"New entry for {artist} {title} ({song['id']} chart {chart})" ) next_id = self.get_next_music_id() else: print( f"Reused entry for {artist} {title} ({song['id']} chart {chart})" ) next_id = old_id self.insert_music_id_for_song( next_id, song["id"], chart, title, artist, genre, { "category": str(song["folder"]), "difficulty": difficulty, }, ) self.finish_batch() class ImportJubeat(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: if version in ["saucer", "saucer-fulfill", "prop", "qubell", "clan", "festo"]: actual_version = { "saucer": VersionConstants.JUBEAT_SAUCER, "saucer-fulfill": VersionConstants.JUBEAT_SAUCER_FULFILL, "prop": VersionConstants.JUBEAT_PROP, "qubell": VersionConstants.JUBEAT_QUBELL, "clan": VersionConstants.JUBEAT_CLAN, "festo": VersionConstants.JUBEAT_FESTO, }.get(version, -1) elif version in ["omni-prop", "omni-qubell", "omni-clan", "omni-festo"]: actual_version = { "omni-prop": VersionConstants.JUBEAT_PROP, "omni-qubell": VersionConstants.JUBEAT_QUBELL, "omni-clan": VersionConstants.JUBEAT_CLAN, "omni-festo": VersionConstants.JUBEAT_FESTO, }.get(version, -1) + DBConstants.OMNIMIX_VERSION_BUMP elif version == "all": actual_version = None if actual_version in [ None, VersionConstants.JUBEAT_FESTO, VersionConstants.JUBEAT_FESTO + DBConstants.OMNIMIX_VERSION_BUMP, ]: # jubeat festo adds in separation of normal and hard mode scores. # This adds a duplicate of each chart so that we show separated scores. self.charts = [0, 1, 2, 3, 4, 5] elif actual_version in [ VersionConstants.JUBEAT_SAUCER, VersionConstants.JUBEAT_SAUCER_FULFILL, VersionConstants.JUBEAT_PROP, VersionConstants.JUBEAT_QUBELL, VersionConstants.JUBEAT_CLAN, VersionConstants.JUBEAT_PROP + DBConstants.OMNIMIX_VERSION_BUMP, VersionConstants.JUBEAT_QUBELL + DBConstants.OMNIMIX_VERSION_BUMP, VersionConstants.JUBEAT_CLAN + DBConstants.OMNIMIX_VERSION_BUMP, ]: self.charts = [0, 1, 2] else: raise CLIException( "Unsupported Jubeat version, expected one of the following: saucer, saucer-fulfill, prop, omni-prop, qubell, omni-qubell, clan, omni-clan, festo, omni-festo!" ) super().__init__( config, GameConstants.JUBEAT, actual_version, no_combine, update ) def scrape(self, xmlfile: str) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]]]: if self.version is None: raise CLIException("Can't scrape Jubeat database for 'all' version!") try: # Probably UTF-8 music DB tree = ET.parse(xmlfile) root = tree.getroot() except ValueError: # Probably shift-jis emblems with open(xmlfile, "rb") as xmlhandle: xmldata = xmlhandle.read().decode("shift_jisx0213") root = ET.fromstring(xmldata) songs: List[Dict[str, Any]] = [] for music_entry in root.find("body") or []: songid = int(music_entry.find("music_id").text) bpm_min = float(music_entry.find("bpm_min").text) bpm_max = float(music_entry.find("bpm_max").text) earliest_version = 0 version = int(music_entry.find("version").text.strip(), 16) while not version & 1: version >>= 1 earliest_version += 1 # Since this is actually 1 to 1, I'm only adding this for clarity rather than necessity version_to_db_constant = { 1: VersionConstants.JUBEAT, 2: VersionConstants.JUBEAT_RIPPLES, 3: VersionConstants.JUBEAT_RIPPLES_APPEND, 4: VersionConstants.JUBEAT_KNIT, 5: VersionConstants.JUBEAT_KNIT_APPEND, 6: VersionConstants.JUBEAT_COPIOUS, 7: VersionConstants.JUBEAT_COPIOUS_APPEND, 8: VersionConstants.JUBEAT_SAUCER, 9: VersionConstants.JUBEAT_SAUCER_FULFILL, 10: VersionConstants.JUBEAT_PROP, 11: VersionConstants.JUBEAT_QUBELL, 12: VersionConstants.JUBEAT_CLAN, 13: VersionConstants.JUBEAT_FESTO, } if bpm_max > 0 and bpm_min < 0: bpm_min = bpm_max if music_entry.find("detail_level_bsc") is not None: difficulties = [ float(music_entry.find("detail_level_bsc").text), float(music_entry.find("detail_level_adv").text), float(music_entry.find("detail_level_ext").text), ] else: difficulties = [ float(music_entry.find("level_bsc").text), float(music_entry.find("level_adv").text), float(music_entry.find("level_ext").text), ] genre = "other" if ( music_entry.find("genre") is not None ): # Qubell extend music_info doesn't have this field for possible_genre in music_entry.find("genre"): if int(possible_genre.text) != 0: genre = str(possible_genre.tag) songs.append( { "id": songid, # Title/artist aren't in the music data for Jubeat and must be manually populated. # This is why there is a separate "import_metadata" and data file. "title": None, "artist": None, "genre": genre, "version": version_to_db_constant.get(earliest_version), "bpm_min": bpm_min, "bpm_max": bpm_max, "difficulty": { "basic": difficulties[0], "advanced": difficulties[1], "extreme": difficulties[2], }, } ) emblems: List[Dict[str, Any]] = [] if self.version in { VersionConstants.JUBEAT_PROP, VersionConstants.JUBEAT_QUBELL, VersionConstants.JUBEAT_CLAN, VersionConstants.JUBEAT_FESTO, }: for emblem_entry in root.find("emblem_list") or []: index = int(emblem_entry.find("index").text) layer = int(emblem_entry.find("layer").text) music_id = int(emblem_entry.find("music_id").text) evolved = int(emblem_entry.find("evolved").text) rarity = int(emblem_entry.find("rarity").text) name = emblem_entry.find("name").text emblems.append( { "id": index, "layer": layer, "music_id": music_id, "evolved": evolved, "rarity": rarity, "name": name, } ) return songs, emblems def lookup( self, server: str, token: str ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]]]: if self.version is None: raise CLIException("Can't look up Jubeat database for 'all' version!") # Grab music info from remote server music = self.remote_music(server, token) songs = music.get_all_songs(self.game, self.version) lut: Dict[int, Dict[str, Any]] = {} chart_map = { 0: "basic", 1: "advanced", 2: "extreme", } # Format it the way we expect for song in songs: if song.chart not in chart_map: # Ignore charts on songs we don't support/care about. continue if song.id not in lut: lut[song.id] = { "id": song.id, "title": song.name, "artist": song.artist, "genre": song.genre, "version": song.data.get_int("version"), "bpm_min": song.data.get_float("bpm_min"), "bpm_max": song.data.get_float("bpm_max"), "difficulty": { "basic": 0.0, "advanced": 0.0, "extreme": 0.0, }, } lut[song.id]["difficulty"][chart_map[song.chart]] = song.data.get_float( "difficulty" ) # Reassemble the data reassembled_songs = [val for _, val in lut.items()] emblems: List[Dict[str, Any]] = [] if self.version in { VersionConstants.JUBEAT_PROP, VersionConstants.JUBEAT_QUBELL, VersionConstants.JUBEAT_CLAN, VersionConstants.JUBEAT_FESTO, }: game = self.remote_game(server, token) for item in game.get_items(self.game, self.version): if item.type == "emblem": emblems.append( { "id": item.id, "layer": item.data.get_int("layer"), "music_id": item.data.get_int("music_id"), "evolved": item.data.get_int("evolved"), "rarity": item.data.get_int("rarity"), "name": item.data.get_str("name"), } ) return reassembled_songs, emblems def __revivals(self, songid: int, chart: int) -> Optional[int]: old_id = self.get_music_id_for_song(songid, chart) if old_id is not None: return old_id # In qubell and clan omnimix, PPAP and Bonjour the world are placed # at this arbitrary songid since they weren't assigned one originally # In jubeat festo, these songs were given proper songids so we need to account for this legacy_to_modern_map = { 71000001: 70000124, # PPAP 71000002: 70000154, # Bonjour the world 50000020: 80000037, # 千本桜 was removed and then revived in clan 60000063: 70000100, # Khamen break sdvx had the first id for prop(never released officially) } modern_to_legacy_map = {v: k for k, v in legacy_to_modern_map.items()} legacy_songid = legacy_to_modern_map.get(songid) if legacy_songid is not None: old_id = self.get_music_id_for_song(legacy_songid, chart) if old_id is not None: return old_id modern_songid = modern_to_legacy_map.get(songid) if modern_songid is not None: old_id = self.get_music_id_for_song(modern_songid, chart) if old_id is not None: return old_id # Failed, so create a new one return None def import_music_db(self, songs: List[Dict[str, Any]]) -> None: if self.version is None: raise CLIException("Can't import Jubeat database for 'all' version!") chart_map: Dict[int, str] = { 0: "basic", 1: "advanced", 2: "extreme", } for song in songs: # Skip over duplicate songs for the "play five different versions of this song # across different prefectures" event. The song ID range is 8000301-8000347, so # we arbitrarily choose to keep only the first one. songid = song["id"] if songid in set(range(80000302, 80000348)): continue self.start_batch() for chart in self.charts: if chart <= 2: # First, try to find in the DB from another version old_id = self.__revivals(songid, chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {songid} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {songid} chart {chart}") next_id = old_id data = { "difficulty": song["difficulty"][chart_map[chart]], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], "version": song["version"], } else: # First, try to find in the DB from another version old_id = self.__revivals(songid, chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {songid} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {songid} chart {chart}") next_id = old_id data = { "difficulty": song["difficulty"][chart_map[chart - 3]], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], "version": song["version"], } self.insert_music_id_for_song( next_id, songid, chart, song["title"], song["artist"], song["genre"], data, ) self.finish_batch() def import_emblems(self, emblems: List[Dict[str, Any]]) -> None: if self.version is None: raise CLIException("Can't import Jubeat database for 'all' version!") self.start_batch() for i, emblem in enumerate(emblems): # Make importing faster but still do it in chunks if (i % 16) == 15: self.finish_batch() self.start_batch() print(f"New catalog entry for {emblem['music_id']}") self.insert_catalog_entry( "emblem", emblem["id"], { "layer": emblem["layer"], "music_id": emblem["music_id"], "evolved": emblem["evolved"], "rarity": emblem["rarity"], "name": emblem["name"], }, ) self.finish_batch() def import_metadata(self, tsvfile: str) -> None: if self.version is not None: raise CLIException( "Unsupported Jubeat version, expected one of the following: all" ) with open(tsvfile, newline="") as tsvhandle: jubeatreader = csv.reader(tsvhandle, delimiter="\t", quotechar='"') for row in jubeatreader: songid = int(row[0]) name = row[1] artist = row[2] print(f"Setting name/artist for {songid} all charts") self.start_batch() for chart in self.charts: self.update_metadata_for_song(songid, chart, name, artist) self.finish_batch() class IIDXScrapeConfiguration: def __init__( self, *, version: str, stride: int, qp_head_offset: int, qp_head_length: int, qp_hair_offset: int, qp_hair_length: int, qp_face_offset: int, qp_face_length: int, qp_hand_offset: int, qp_hand_length: int, qp_body_offset: int, qp_body_length: int, filename_offset: int, qpro_id_offset: Optional[int], packedfmt: str, ) -> None: self.version = version self.stride = stride self.qp_head_offset = qp_head_offset self.qp_head_length = qp_head_length self.qp_hair_offset = qp_hair_offset self.qp_hair_length = qp_hair_length self.qp_face_offset = qp_face_offset self.qp_face_length = qp_face_length self.qp_hand_offset = qp_hand_offset self.qp_hand_length = qp_hand_length self.qp_body_offset = qp_body_offset self.qp_body_length = qp_body_length self.filename_offset = filename_offset self.qpro_id_offset = qpro_id_offset self.packedfmt = packedfmt class ImportIIDX(ImportBase): # Tutorial charts that shouldn't be on the UI BANNED_CHARTS = [ 16070, 16071, 16072, 16080, 16081, 16082, ] def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: if version in ["20", "21", "22", "23", "24", "25", "26"]: actual_version = { "20": VersionConstants.IIDX_TRICORO, "21": VersionConstants.IIDX_SPADA, "22": VersionConstants.IIDX_PENDUAL, "23": VersionConstants.IIDX_COPULA, "24": VersionConstants.IIDX_SINOBUZ, "25": VersionConstants.IIDX_CANNON_BALLERS, "26": VersionConstants.IIDX_ROOTAGE, }[version] self.charts = [0, 1, 2, 3, 4, 5, 6] elif version in [ "omni-20", "omni-21", "omni-22", "omni-23", "omni-24", "omni-25", "omni-26", ]: actual_version = { "omni-20": VersionConstants.IIDX_TRICORO, "omni-21": VersionConstants.IIDX_SPADA, "omni-22": VersionConstants.IIDX_PENDUAL, "omni-23": VersionConstants.IIDX_COPULA, "omni-24": VersionConstants.IIDX_SINOBUZ, "omni-25": VersionConstants.IIDX_CANNON_BALLERS, "omni-26": VersionConstants.IIDX_ROOTAGE, }[version] + DBConstants.OMNIMIX_VERSION_BUMP self.charts = [0, 1, 2, 3, 4, 5, 6] elif version == "all": actual_version = None self.charts = [0, 1, 2, 3, 4, 5, 6] else: raise CLIException( "Unsupported IIDX version, expected one of the following: 20, 21, 22, 23, 24, 25, 26, omni-20, omni-21, omni-22, omni-23, omni-24, omni-25, omni-26!" ) super().__init__(config, GameConstants.IIDX, actual_version, no_combine, update) def __gather_sound_files(self, directory: str) -> Dict[int, str]: files = {} for dirpath, dirnames, filenames in os.walk(directory): for filename in filenames: songid, extension = os.path.splitext(filename) if extension == ".1" or extension == ".ifs": try: files[int(songid)] = os.path.join( directory, os.path.join(dirpath, filename) ) except ValueError: # Invalid file pass for dirname in dirnames: files.update( self.__gather_sound_files(os.path.join(directory, dirname)) ) return files def __revivals(self, songid: int, chart: int) -> Optional[int]: old_id = self.get_music_id_for_song(songid, chart) if old_id is not None: return old_id # For revivals from older games, these show up as their respective old IDs # in Spada Omnimix, but in Pendual Omnimix they're in the Pendual category. legacy_to_modern_map = { 4213: 23066, 9203: 22068, 10203: 22052, 12201: 22039, 12204: 21201, 12206: 21064, 13215: 23077, 14202: 22025, 14210: 21068, 14211: 22069, 14214: 23070, 15202: 23069, 15204: 21063, 15205: 21065, 15207: 22028, 15208: 22049, 15209: 22043, 15211: 23060, 15215: 21062, 16207: 21067, 16209: 23062, 16212: 21066, 22096: 23030, 22097: 23051, 21214: 11101, 21221: 14101, 21225: 15104, 21226: 15102, 21231: 15101, 21237: 15103, 21240: 16105, 21242: 16104, 21253: 16103, 21258: 16102, 21262: 16101, 21220: 14100, } # Some charts were changed, and others kept the same on these if chart in [0, 1, 2]: legacy_to_modern_map[9206] = 23065 legacy_songid = legacy_to_modern_map.get(songid) if legacy_songid is not None: old_id = self.get_music_id_for_song(legacy_songid, chart) if old_id is not None: return old_id modern_to_legacy_map = { 23066: 4213, 22068: 9203, 22052: 10203, 22039: 12201, 21201: 12204, 21064: 12206, 23077: 13215, 22025: 14202, 21068: 14210, 22069: 14211, 23070: 14214, 23069: 15202, 21063: 15204, 21065: 15205, 22028: 15207, 22049: 15208, 22043: 15209, 23060: 15211, 21062: 15215, 21067: 16207, 23062: 16209, 21066: 16212, 23030: 22096, 23051: 22097, 11101: 21214, 14101: 21221, 15104: 21225, 15102: 21226, 15101: 21231, 15103: 21237, 16105: 21240, 16104: 21242, 16103: 21253, 16102: 21258, 16101: 21262, 14100: 21220, } # Some charts were changed, and others kept the same on tehse if chart in [0, 1, 2]: modern_to_legacy_map[23065] = 9206 modern_songid = modern_to_legacy_map.get(songid) if modern_songid is not None: old_id = self.get_music_id_for_song(modern_songid, chart) if old_id is not None: return old_id # Failed, so create a new one return None def __charts(self, songid: int, chart: int) -> int: # Scripted connection long was set as a hyper in Tricoro omnimix, we # need to map the charts to the another in every other version. if songid == 12204: if chart == 1: return 2 if chart == 2: return 1 return chart def scrape( self, binfile: str, assets_dir: Optional[str] ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]]]: if self.version is None: raise CLIException("Can't import IIDX database for 'all' version!") if assets_dir is not None: sound_files = self.__gather_sound_files(os.path.abspath(assets_dir)) else: sound_files = None bh = open(binfile, "rb") try: binarydata = bh.read() finally: bh.close() import_qpros = True # by default, try to import qpros try: pe = PEFile(binarydata) except BaseException: import_qpros = False # if it failed then we're reading a music db file, not the executable if not import_qpros: # We should assume this is a standard music_data.bin, and read the soungs out of it. songs: List[Dict[str, Any]] = [] musicdb = IIDXMusicDB(binarydata) for song in musicdb.songs: bpm = (0, 0) notecounts = [0, 0, 0, 0, 0, 0] if song.id in self.BANNED_CHARTS: continue if sound_files is not None: if song.id in sound_files: # Look up chart info! filename = sound_files[song.id] _, extension = os.path.splitext(filename) data = None if extension == ".1": fp = open(filename, "rb") data = fp.read() fp.close() else: fp = open(filename, "rb") ifsdata = fp.read() fp.close() ifs = IFS(ifsdata) for fn in ifs.filenames: _, extension = os.path.splitext(fn) if extension == ".1": data = ifs.read_file(fn) if data is not None: iidxchart = IIDXChart(data) bpm_min, bpm_max = iidxchart.bpm bpm = (bpm_min, bpm_max) notecounts = iidxchart.notecounts else: print( f"Could not find chart information for song {song.id}!" ) else: print( f"No chart information because chart for song {song.id} is missing!" ) songs.append( { "id": song.id, "title": song.title, "artist": song.artist, "genre": song.genre, "bpm_min": bpm[0], "bpm_max": bpm[1], "difficulty": { "spn": song.difficulties[0], "sph": song.difficulties[1], "spa": song.difficulties[2], "dpn": song.difficulties[3], "dph": song.difficulties[4], "dpa": song.difficulties[5], }, "notecount": { "spn": notecounts[0], "sph": notecounts[1], "spa": notecounts[2], "dpn": notecounts[3], "dph": notecounts[4], "dpa": notecounts[5], }, } ) # We only import one or the other here, I know its a weird function. return songs, [] else: # We should assume that we want to extract QPro data from this DLL. configurations: List[IIDXScrapeConfiguration] = [] if self.version == VersionConstants.IIDX_TRICORO: # Based on LDJ:J:A:A:2013090900 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2013090900", stride=4, qp_head_offset=0x1CCB18, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=79, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x1CCC58, qp_hair_length=103, qp_face_offset=0x1CCDF8, qp_face_length=50, qp_hand_offset=0x1CCEC0, qp_hand_length=103, qp_body_offset=0x1CD060, qp_body_length=106, filename_offset=0, qpro_id_offset=None, packedfmt="I", # filename ) ) if self.version == VersionConstants.IIDX_SPADA: # Based on LDJ:J:A:A:2014071600 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2014071600", stride=4, qp_head_offset=0x213B50, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=125, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x213D48, qp_hair_length=126, qp_face_offset=0x213F40, qp_face_length=72, qp_hand_offset=0x214060, qp_hand_length=135, qp_body_offset=0x214280, qp_body_length=135, filename_offset=0, qpro_id_offset=None, packedfmt="I", # filename ) ) if self.version == VersionConstants.IIDX_PENDUAL: # Based on LDJ:J:A:A:2015080500 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2015080500", stride=4, qp_head_offset=0x1D5228, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=163, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x1D54B8, qp_hair_length=182, qp_face_offset=0x1D5790, qp_face_length=106, qp_hand_offset=0x1D5938, qp_hand_length=184, qp_body_offset=0x1D5C18, qp_body_length=191, filename_offset=0, qpro_id_offset=None, packedfmt="I", # filename ) ) if self.version == VersionConstants.IIDX_COPULA: # Based on LDJ:J:A:A:2016083100 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2016083100", stride=8, qp_head_offset=0x12F9D8, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=186, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x12FFA8, qp_hair_length=202, qp_face_offset=0x1305F8, qp_face_length=126, qp_hand_offset=0x1309E8, qp_hand_length=206, qp_body_offset=0x131058, qp_body_length=211, filename_offset=0, qpro_id_offset=1, packedfmt="I" "I", # filename # string containing id and name of the part ) ) if self.version == VersionConstants.IIDX_SINOBUZ: # Based on LDJ:J:A:A:2017082800 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2017082800", stride=8, qp_head_offset=0x149F88, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=211, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x14A620, qp_hair_length=245, qp_face_offset=0x14ADC8, qp_face_length=152, qp_hand_offset=0x14B288, qp_hand_length=236, qp_body_offset=0x14B9E8, qp_body_length=256, filename_offset=0, qpro_id_offset=1, packedfmt="I" "I", # filename # string containing id and name of the part ) ) if self.version == VersionConstants.IIDX_CANNON_BALLERS: # Based on LDJ:J:A:A:2018091900 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2018091900", stride=16, qp_head_offset=0x2339E0, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=231, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x234850, qp_hair_length=267, qp_face_offset=0x235900, qp_face_length=173, qp_hand_offset=0x2363D0, qp_hand_length=261, qp_body_offset=0x237420, qp_body_length=282, filename_offset=0, qpro_id_offset=1, packedfmt="Q" "Q", # filename # string containing id and name of the part ) ) if self.version == VersionConstants.IIDX_ROOTAGE: # Based on LDJ:J:A:A:2019090200 configurations.append( IIDXScrapeConfiguration( version="LDJ:J:A:A:2019090200", stride=16, qp_head_offset=0x5065F0, # qpro body parts are stored in 5 separate arrays in the game data, since there can be collision in qp_head_length=259, # the qpro id numbers, it's best to store them as separate types in the catalog as well. qp_hair_offset=0x507620, qp_hair_length=288, qp_face_offset=0x508820, qp_face_length=193, qp_hand_offset=0x509430, qp_hand_length=287, qp_body_offset=0x50A620, qp_body_length=304, filename_offset=0, qpro_id_offset=1, packedfmt="Q" "Q", # filename # string containing id and name of the part ) ) def read_string(offset: int) -> str: # First, translate load offset in memory to disk offset offset = pe.virtual_to_physical(offset) # Now, grab bytes until we're null-terminated bytestring = [] while binarydata[offset] != 0: bytestring.append(binarydata[offset]) offset = offset + 1 # Its shift-jis encoded, so decode it now return bytes(bytestring).decode("shift_jisx0213") def read_qpro_db( qpros: List[Dict[str, Any]], config: IIDXScrapeConfiguration, offset: int, length: int, qp_type: str, ) -> None: for qpro_id in range(length): chunkoffset = offset + (config.stride * qpro_id) chunkdata = binarydata[chunkoffset : (chunkoffset + config.stride)] unpacked = struct.unpack(config.packedfmt, chunkdata) filename = read_string(unpacked[config.filename_offset]).replace( "qp_", "" ) remove = f"_{qp_type}.ifs" filename = ( filename.replace(remove, "") .replace("_head1.ifs", "") .replace("_head2.ifs", "") ) if config.qpro_id_offset is None: name = filename # qpro names are not stored in these games so use the identifier instead else: name = read_string(unpacked[config.qpro_id_offset])[ 4: ] # qpro name is stored in second string of form "000:name" qproinfo = { "identifier": filename, "id": qpro_id, "name": name, "type": qp_type, } qpros.append(qproinfo) for config in configurations: try: print(f"Trying configuration for game version {config.version}...") qpros: List[Dict[str, Any]] = [] # For each category, attempt to append all the types. read_qpro_db( qpros, config, config.qp_head_offset, config.qp_head_length, "head", ) read_qpro_db( qpros, config, config.qp_hair_offset, config.qp_hair_length, "hair", ) read_qpro_db( qpros, config, config.qp_face_offset, config.qp_face_length, "face", ) read_qpro_db( qpros, config, config.qp_hand_offset, config.qp_hand_length, "hand", ) read_qpro_db( qpros, config, config.qp_body_offset, config.qp_body_length, "body", ) # If we got here, that means we ran into no issues and didn't have to attempt another offset. print("Successfully parsed game DB!") # We only import one or the other here, I know its a weird function. return [], qpros except (UnicodeError, InvalidOffsetException): # These offsets are possibly not correct, so try the next configuration. print( "Failed to parse game DB using current inferred data version!" ) pass raise CLIException( f"Could not determine correct binary parser configuration for IIDX version {self.version}" ) def lookup( self, server: str, token: str ) -> Tuple[List[Dict[str, Any]], List[Dict[str, Any]]]: if self.version is None: raise CLIException("Can't look up IIDX database for 'all' version!") # Grab music info from remote server music = self.remote_music(server, token) songs = music.get_all_songs(self.game, self.version) lut: Dict[int, Dict[str, Any]] = {} chart_map = { 0: "spn", 1: "sph", 2: "spa", 3: "dpn", 4: "dph", 5: "dpa", } # Format it the way we expect for song in songs: if song.id in self.BANNED_CHARTS: continue if song.chart not in chart_map: # Ignore charts on songs we don't support/care about. continue if song.id not in lut: lut[song.id] = { "id": song.id, "title": song.name, "artist": song.artist, "genre": song.genre, "bpm_min": song.data.get_int("bpm_min"), "bpm_max": song.data.get_int("bpm_max"), "difficulty": { "spn": 0, "sph": 0, "spa": 0, "dpn": 0, "dph": 0, "dpa": 0, }, "notecount": { "spn": 0, "sph": 0, "spa": 0, "dpn": 0, "dph": 0, "dpa": 0, }, } if song.chart in chart_map: lut[song.id]["difficulty"][chart_map[song.chart]] = song.data.get_int( "difficulty" ) lut[song.id]["notecount"][chart_map[song.chart]] = song.data.get_int( "notecount" ) # Return the reassembled data qpros: List[Dict[str, Any]] = [] game = self.remote_game(server, token) for item in game.get_items(self.game, self.version): if "qp_" in item.type: qpros.append( { "identifier": item.data.get_str("identifier"), "id": item.id, "name": item.data.get_str("name"), "type": item.data.get_str("type"), } ) return [val for _, val in lut.items()], qpros def import_music_db(self, songs: List[Dict[str, Any]]) -> None: if self.version is None: raise CLIException("Can't import IIDX database for 'all' version!") # Import each song into our DB chart_map = { 0: "spn", 1: "sph", 2: "spa", 3: "dpn", 4: "dph", 5: "dpa", } for song in songs: self.start_batch() for chart in self.charts: if chart == 6: # Beginner chart songdata: Dict[str, Any] = {} else: songdata = { "difficulty": song["difficulty"][chart_map[chart]], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], "notecount": song["notecount"][chart_map[chart]], } # First, try to find in the DB from another version old_id = self.__revivals(song["id"], self.__charts(song["id"], chart)) if self.no_combine or old_id is None: # Insert original print(f"New entry for {song['id']} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {song['id']} chart {chart}") next_id = old_id self.insert_music_id_for_song( next_id, song["id"], chart, song["title"], song["artist"], song["genre"], songdata, ) self.finish_batch() def import_qpros(self, qpros: List[Dict[str, Any]]) -> None: if self.version is None: raise CLIException("Can't import IIDX database for 'all' version!") self.start_batch() for i, qpro in enumerate(qpros): # Make importing faster but still do it in chunks if (i % 16) == 15: self.finish_batch() self.start_batch() print(f"New catalog entry for {qpro['id']}") self.insert_catalog_entry( f"qp_{qpro['type']}", qpro["id"], { "name": qpro["name"], "identifier": qpro["identifier"], }, ) self.finish_batch() def import_metadata(self, tsvfile: str) -> None: if self.version is not None: raise CLIException( "Unsupported IIDX version, expected one of the following: all" ) with open(tsvfile, newline="") as tsvhandle: iidxreader = csv.reader(tsvhandle, delimiter="\t", quotechar='"') for row in iidxreader: songid = int(row[0]) name = row[1] artist = row[2] genre = row[3] if len(name) == 0: name = None if len(artist) == 0: artist = None if len(genre) == 0: genre = None print(f"Setting name/artist/genre for {songid} all charts") self.start_batch() for chart in self.charts: self.update_metadata_for_song(songid, chart, name, artist, genre) self.finish_batch() class DDRScrapeConfiguration: def __init__( self, *, version: str, offset: int, size: int, length: int, unpackfmt: str, id_offset: int, edit_offset: int, bpm_min_offset: int, bpm_max_offset: int, folder_offset: int, single_difficulties: int, double_difficulties: int, groove_single_beginner: int, groove_single_basic: int, groove_single_difficult: int, groove_single_expert: int, groove_single_challenge: int, groove_double_basic: int, groove_double_difficult: int, groove_double_expert: int, groove_double_challenge: int, voltage: int, stream: int, air: int, chaos: int, freeze: int, folder_start: int, ) -> None: self.version = version self.offset = offset self.size = size self.length = length self.unpackfmt = unpackfmt self.id_offset = id_offset self.edit_offset = edit_offset self.bpm_min_offset = bpm_min_offset self.bpm_max_offset = bpm_max_offset self.folder_offset = folder_offset self.single_difficulties = single_difficulties self.double_difficulties = double_difficulties self.groove_single_beginner = groove_single_beginner self.groove_single_basic = groove_single_basic self.groove_single_difficult = groove_single_difficult self.groove_single_expert = groove_single_expert self.groove_single_challenge = groove_single_challenge self.groove_double_basic = groove_double_basic self.groove_double_difficult = groove_double_difficult self.groove_double_expert = groove_double_expert self.groove_double_challenge = groove_double_challenge self.voltage = voltage self.stream = stream self.air = air self.chaos = chaos self.freeze = freeze self.folder_start = folder_start class ImportDDR(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: if version in ["12", "13", "14", "15", "16"]: actual_version = { "12": VersionConstants.DDR_X2, "13": VersionConstants.DDR_X3_VS_2NDMIX, "14": VersionConstants.DDR_2013, "15": VersionConstants.DDR_2014, "16": VersionConstants.DDR_ACE, }[version] self.charts = [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] else: raise CLIException( "Unsupported DDR version, expected one of the following: 12, 13, 14, 15, 16" ) super().__init__(config, GameConstants.DDR, actual_version, no_combine, update) def scrape(self, infile: str) -> List[Dict[str, Any]]: with open(infile, mode="rb") as myfile: data = myfile.read() myfile.close() def add_skew(unpackfmt: str, size: int, skew: int) -> str: # Skew is because I'm too lazy to count the Hs in the format, so just # pad it for ease of construction here. return unpackfmt + ("x" * (size - len(unpackfmt) - skew)) configurations: List[DDRScrapeConfiguration] = [] if self.version == VersionConstants.DDR_X2: # Based on JDX:J:A:A:2010111000 configurations.append( DDRScrapeConfiguration( version="JDX:J:A:A:2010111000", offset=0x254FC0, size=0x14C, length=894, unpackfmt=add_skew( ( # Basic stuff like ID, bpm, chart difficulties " List[Dict[str, Any]]: tree = ET.parse(infile) root = tree.getroot() data = {} for music_entry in root.find("mdblist"): musicid = int(music_entry.attrib["reclink"]) title = "" artist = "" title_en = "" artist_en = "" for child in music_entry: if child.tag == "name": if child.attrib["lang"] == "ja": title = child.text else: title_en = child.text if child.tag == "artist": if child.attrib["lang"] == "ja": artist = child.text else: artist_en = child.text data[musicid] = { "title": title, "artist": artist, "title_en": title_en, "artist_en": artist_en, } songs = copy.deepcopy(songs) for song in songs: newdata = data.get(song["id"]) if newdata is not None: song.update(newdata) return songs def parse_xml(self, arcfile: str) -> List[Dict[str, Any]]: with open(arcfile, mode="rb") as myfile: data = myfile.read() myfile.close() arc = ARC(data) if "data/gamedata/musicdb.xml" in arc.filenames: xmldata = arc.read_file("data/gamedata/musicdb.xml") else: raise CLIException("Invalid .arc file provided!") xml = xmldata.decode("utf-8") root = ET.fromstring(xml) songs = [] if root.tag != "mdb": raise Exception("Invalid musicdb.xml file in .arc!") for music_entry in root: songid = int(music_entry.find("mcode").text) title = music_entry.find("title").text artist = music_entry.find("artist").text ssqcode = music_entry.find("basename").text bpm = int(music_entry.find("bpmmax").text) folder = int(music_entry.find("series").text) difficulties = [int(x) for x in music_entry.find("diffLv").text.split(" ")] # For some reason Ace thinks of itself as 17, and DDR 2013/2014 as 14, 15 and 16 # somewhat spread out. Fix that here. folder = { 1: 1, 2: 2, 3: 3, 4: 4, 5: 5, 6: 6, 7: 7, 8: 8, 9: 9, 10: 10, 11: 11, 12: 12, 13: 13, 14: 14, 15: 15, 16: 15, 17: 16, 18: 17, }[folder] songinfo = { "id": songid, "edit_id": songid, "ssqcode": ssqcode, "title": title, "artist": artist, "difficulty": { "single": { "beginner": difficulties[0], "basic": difficulties[1], "difficult": difficulties[2], "expert": difficulties[3], "challenge": difficulties[4], }, "double": { "beginner": difficulties[5], "basic": difficulties[6], "difficult": difficulties[7], "expert": difficulties[8], "challenge": difficulties[9], }, }, "bpm_min": bpm, "bpm_max": bpm, "folder": folder, "groove_gauge": {}, } # Groove information is calculated on the fly, maybe some day we will # duplicate that here, but for now, zero it out. for playmode in ["single", "double"]: songinfo["groove_gauge"][playmode] = {} # type: ignore for charttype in [ "beginner", "basic", "difficult", "expert", "challenge", ]: songinfo["groove_gauge"][playmode][charttype] = { # type: ignore "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, } songs.append(songinfo) return songs def lookup(self, server: str, token: str) -> List[Dict[str, Any]]: # Grab music info from remote server music = self.remote_music(server, token) songs = music.get_all_songs(self.game, self.version) lut: Dict[int, Dict[str, Any]] = {} chart_map = { 0: ("single", "beginner"), 1: ("single", "basic"), 2: ("single", "difficult"), 3: ("single", "expert"), 4: ("single", "challenge"), 5: ("double", "beginner"), 6: ("double", "basic"), 7: ("double", "difficult"), 8: ("double", "expert"), 9: ("double", "challenge"), } # Format it the way we expect for song in songs: if song.chart not in chart_map: # Ignore charts on songs we don't support/care about. continue if song.id not in lut: lut[song.id] = { "id": song.id, "edit_id": song.data.get_int("edit_id"), "ssqcode": "", "title": song.name, "artist": song.artist, "difficulty": { "single": { "beginner": 0, "basic": 0, "difficult": 0, "expert": 0, "challenge": 0, }, "double": { "beginner": 0, "basic": 0, "difficult": 0, "expert": 0, "challenge": 0, }, }, "groove_gauge": { "single": { "beginner": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "basic": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "difficult": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "expert": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "challenge": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, }, "double": { "beginner": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "basic": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "difficult": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "expert": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, "challenge": { "voltage": 0, "stream": 0, "air": 0, "chaos": 0, "freeze": 0, }, }, }, "bpm_min": song.data.get_int("bpm_min"), "bpm_max": song.data.get_int("bpm_max"), "folder": song.data.get_int("category"), } style, chart = chart_map[song.chart] lut[song.id]["difficulty"][style][chart] = song.data.get_int("difficulty") lut[song.id]["groove_gauge"][style][chart]["air"] = song.data.get_dict( "groove" ).get_int("air") lut[song.id]["groove_gauge"][style][chart]["chaos"] = song.data.get_dict( "groove" ).get_int("chaos") lut[song.id]["groove_gauge"][style][chart]["freeze"] = song.data.get_dict( "groove" ).get_int("freeze") lut[song.id]["groove_gauge"][style][chart]["stream"] = song.data.get_dict( "groove" ).get_int("stream") lut[song.id]["groove_gauge"][style][chart]["voltage"] = song.data.get_dict( "groove" ).get_int("voltage") # Return the reassembled data return [val for _, val in lut.items()] def import_music_db(self, songs: List[Dict[str, Any]]) -> None: for song in songs: self.start_batch() for chart in self.charts: key = ["beginner", "basic", "difficult", "expert", "challenge"] if chart in [0, 1, 2, 3, 4]: # For singles only! difficulty = song["difficulty"]["single"][key[chart]] groovestats = song["groove_gauge"]["single"][key[chart]] elif chart in [5, 6, 7, 8, 9]: # Doubles! difficulty = song["difficulty"]["double"][key[chart - 5]] groovestats = song["groove_gauge"]["double"][key[chart - 5]] else: raise Exception("Unrecognized chart type!") if difficulty == 0: # No chart for this difficulty continue if song["edit_id"] == 0: raise Exception("Expected non-zero edit id!") # DDR is stupid and changes in-game IDs around willy-nilly, but the edit ID is stable. # So, create a virtual edit ID entry and link everything to that. We can't just store # the edit ID as the real ID because in-game the protocol uses the changing ID. old_id = self.get_music_id_for_song(song["edit_id"], chart, version=0) if self.no_combine or old_id is None: # Insert original print( f"New entry for {song['title']} {song['artist']} ({song['id']} chart {chart})" ) next_id = self.get_next_music_id() else: print( f"Reused entry for {song['title']} {song['artist']} ({song['id']} chart {chart})" ) next_id = old_id # Add the virtual entry we talked about above, so we can link this song in the future. self.insert_music_id_for_song( next_id, song["edit_id"], chart, song["title"], song["artist"], None, # No genres in DDR { "category": song["folder"], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], "difficulty": difficulty, "groove": groovestats, "edit_id": song["edit_id"], }, version=0, ) # Add the normal entry so the game finds the song. self.insert_music_id_for_song( next_id, song["id"], chart, song["title"], song["artist"], None, # No genres in DDR { "category": song["folder"], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], "difficulty": difficulty, "groove": groovestats, "edit_id": song["edit_id"], }, ) self.finish_batch() class ImportSDVX(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: actual_version = { "1": VersionConstants.SDVX_BOOTH, "2": VersionConstants.SDVX_INFINITE_INFECTION, "3": VersionConstants.SDVX_GRAVITY_WARS, "4": VersionConstants.SDVX_HEAVENLY_HAVEN, }.get(version, -1) if actual_version == VersionConstants.SDVX_BOOTH: self.charts = [0, 1, 2] elif actual_version in [ VersionConstants.SDVX_INFINITE_INFECTION, VersionConstants.SDVX_GRAVITY_WARS, ]: self.charts = [0, 1, 2, 3] elif actual_version == VersionConstants.SDVX_HEAVENLY_HAVEN: self.charts = [0, 1, 2, 3, 4] else: raise CLIException( "Unsupported SDVX version, expected one of the following: 1, 2, 3, 4!" ) super().__init__(config, GameConstants.SDVX, actual_version, no_combine, update) def scrape(self, infile: str) -> List[Dict[str, Any]]: with open(infile, mode="rb") as myfile: data = myfile.read() myfile.close() pe = PEFile(data) # Ideally we would have multiple configurations supported here, like we do for DDR, # Pop'n, IIDX QPros and Reflec Beat. However, the chances that somebody is trying to # import non-final SDVX Booth catalog items is low enough that it doesn't seem worth # it to try to support this. if self.version == VersionConstants.SDVX_BOOTH: offset = 0xFFF28 size = 163 stride = 40 else: raise CLIException("Unsupported version {self.version} for catalog scrape!") def read_string(spot: int) -> str: # First, translate load offset in memory to disk offset spot = pe.virtual_to_physical(spot) # Now, grab bytes until we're null-terminated bytestring = [] while data[spot] != 0: bytestring.append(data[spot]) spot = spot + 1 return bytes(bytestring).decode("shift_jis") entries = [] for i in range(size): start = offset + i * stride end = offset + (i + 1) * stride chunk = data[start:end] values = struct.unpack(" None: entries = self.scrape(dllfile) for entry in entries: self.start_batch() print(f"New catalog entry for {entry['musicid']} chart {entry['chart']}") self.insert_catalog_entry( "song_unlock", entry["catalogid"], { "blocks": entry["price"], "musicid": entry["musicid"], "chart": entry["chart"], }, ) self.finish_batch() def import_appeal_cards(self, csvfile: str) -> None: with open(csvfile, "rb") as csvhandle: csvdata = csvhandle.read().decode("shift_jisx0213") csvstr = io.StringIO(csvdata) appealreader = csv.reader(csvstr, delimiter=",", quotechar='"') for row in appealreader: appealids = [] if self.version == VersionConstants.SDVX_INFINITE_INFECTION: try: appealids.append(int(row[-5])) except (TypeError, ValueError): pass elif self.version == VersionConstants.SDVX_GRAVITY_WARS: try: appealids.append(int(row[-9])) except (TypeError, ValueError): pass else: raise CLIException( f"Cannot import appeal cards for SDVX version {self.version}" ) self.start_batch() for appealid in appealids: print(f"New catalog entry for appeal card {appealid}") self.insert_catalog_entry( "appealcard", appealid, {}, ) self.finish_batch() def import_music_db_or_appeal_cards(self, xmlfile: str) -> None: with open(xmlfile, "rb") as fp: # This is gross, but elemtree won't do it for us so whatever bytedata = fp.read() strdata = bytedata.decode("shift_jisx0213", errors="replace") root = ET.fromstring(strdata) for music_entry in root.findall("music"): # Grab the ID songid = int(music_entry.attrib["id"]) title = None artist = None bpm_min = None bpm_max = None limited = [0, 0, 0, 0, 0] difficulties = [0, 0, 0, 0, 0] if self.version == VersionConstants.SDVX_BOOTH: # Find normal info about the song for info in music_entry.findall("info"): if info.attrib["attr"] == "title_yomigana": title = jaconv.h2z(info.text) if info.attrib["attr"] == "artist_yomigana": artist = jaconv.h2z(info.text) if info.attrib["attr"] == "bpm_min": bpm_min = float(info.text) if info.attrib["attr"] == "bpm_max": bpm_max = float(info.text) if info.attrib["attr"] == "limited": limited = [ int(info.text), int(info.text), int(info.text), int(info.text), ] # Make sure we got everything if ( title is None or artist is None or bpm_min is None or bpm_max is None ): raise Exception(f"Couldn't parse info for song {songid}") # Grab valid difficulties for difficulty in music_entry.findall("difficulty"): # Figure out the actual difficulty offset = { "novice": 0, "advanced": 1, "exhaust": 2, }.get(difficulty.attrib["attr"]) if offset is None: continue difficulties[offset] = int(difficulty.find("difnum").text) elif self.version in [ VersionConstants.SDVX_INFINITE_INFECTION, VersionConstants.SDVX_GRAVITY_WARS, ]: # Find normal info about the song info = music_entry.find("info") title = info.find("title_name").text artist = info.find("artist_name").text bpm_min = float(info.find("bpm_min").text) / 100.0 bpm_max = float(info.find("bpm_max").text) / 100.0 # Grab valid difficulties for difficulty in music_entry.find("difficulty"): # Figure out the actual difficulty offset = { "novice": 0, "advanced": 1, "exhaust": 2, "infinite": 3, }.get(difficulty.tag) if offset is None: continue difficulties[offset] = int(difficulty.find("difnum").text) limited[offset] = int(difficulty.find("limited").text) elif self.version == VersionConstants.SDVX_HEAVENLY_HAVEN: # Find normal info about the song info = music_entry.find("info") title = info.find("title_name").text artist = info.find("artist_name").text bpm_min = float(info.find("bpm_min").text) / 100.0 bpm_max = float(info.find("bpm_max").text) / 100.0 # Grab valid difficulties for difficulty in music_entry.find("difficulty"): # Figure out the actual difficulty offset = { "novice": 0, "advanced": 1, "exhaust": 2, "infinite": 3, "maximum": 4, }.get(difficulty.tag) if offset is None: continue difficulties[offset] = int(difficulty.find("difnum").text) limited[offset] = int(difficulty.find("limited").text) # Fix accent issues with title/artist accent_lut: Dict[str, str] = { "驩": "Ø", "齲": "♥", "齶": "♡", "趁": "Ǣ", "騫": "á", "曦": "à", "驫": "ā", "齷": "é", "曩": "è", "䧺": "ê", "骭": "ü", } for orig, rep in accent_lut.items(): title = title.replace(orig, rep) artist = artist.replace(orig, rep) # Import it self.start_batch() for chart in self.charts: # First, try to find in the DB from another version old_id = self.get_music_id_for_song(songid, chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {songid} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {songid} chart {chart}") next_id = old_id data = { "limited": limited[chart], "difficulty": difficulties[chart], "bpm_min": bpm_min, "bpm_max": bpm_max, } self.insert_music_id_for_song( next_id, songid, chart, title, artist, None, data ) self.finish_batch() appealids: List[int] = [] for appeal_entry in root.findall("card"): # Grab the ID appealids.append(int(appeal_entry.attrib["id"])) if appealids: self.start_batch() for appealid in appealids: print(f"New catalog entry for appeal card {appealid}") self.insert_catalog_entry( "appealcard", appealid, {}, ) self.finish_batch() def import_from_server(self, server: str, token: str) -> None: # First things first, lets try to import the music DB. We want to make # sure that even if the server doesn't respond right, we have a chart # entry for every chart for each song we're importing. music = self.remote_music(server, token) music_lut: Dict[int, Dict[int, Song]] = {} for entry in music.get_all_songs(self.game, self.version): if entry.id not in music_lut: music_lut[entry.id] = { chart: Song( entry.game, entry.version, entry.id, chart, entry.name, entry.artist, entry.genre, {}, ) for chart in self.charts } music_lut[entry.id][entry.chart] = entry # Import it for _, songs in music_lut.items(): self.start_batch() for _, song in songs.items(): old_id = self.get_music_id_for_song(song.id, song.chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {song.id} chart {song.chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {song.id} chart {song.chart}") next_id = old_id data = { "limited": song.data.get_int("limited"), "difficulty": song.data.get_int("difficulty"), "bpm_min": song.data.get_int("bpm_min"), "bpm_max": song.data.get_int("bpm_max"), } self.insert_music_id_for_song( next_id, song.id, song.chart, song.name, song.artist, None, data ) self.finish_batch() # Now, attempt to insert any catalog items we got for this version. game = self.remote_game(server, token) self.start_batch() for item in game.get_items(self.game, self.version): if item.type == "appealcard": print(f"New catalog entry for appeal card {item.id}") self.insert_catalog_entry( "appealcard", item.id, {}, ) elif item.type == "song_unlock": print( f"New catalog entry for {item.data.get_int('musicid')} chart {item.data.get_int('chart')}" ) self.insert_catalog_entry( "song_unlock", item.id, { "blocks": item.data.get_int("blocks"), "musicid": item.data.get_int("musicid"), "chart": item.data.get_int("chart"), }, ) self.finish_batch() class ImportMuseca(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: if version in ["1", "1+1/2", "plus"]: actual_version = { "1": VersionConstants.MUSECA, "1+1/2": VersionConstants.MUSECA_1_PLUS, "plus": VersionConstants.MUSECA_1_PLUS + DBConstants.OMNIMIX_VERSION_BUMP, }.get(version, -1) if actual_version in [ VersionConstants.MUSECA, VersionConstants.MUSECA_1_PLUS, VersionConstants.MUSECA_1_PLUS + DBConstants.OMNIMIX_VERSION_BUMP, ]: self.charts = [0, 1, 2, 3] else: raise CLIException( "Unsupported Museca version, expected one of the following: 1, 1+1/2, plus!" ) super().__init__( config, GameConstants.MUSECA, actual_version, no_combine, update ) def import_music_db(self, xmlfile: str) -> None: with open(xmlfile, "rb") as fp: # This is gross, but elemtree won't do it for us so whatever bytedata = fp.read() strdata = bytedata.decode("shift_jisx0213") root = ET.fromstring(strdata) for music_entry in root.findall("music"): # Grab the ID songid = int(music_entry.attrib["id"]) title = None artist = None bpm_min = None bpm_max = None limited = [0, 0, 0, 0] difficulties = [0, 0, 0, 0] # Find normal info about the song info = music_entry.find("info") title = info.find("title_name").text artist = info.find("artist_name").text bpm_min = float(info.find("bpm_min").text) / 100.0 bpm_max = float(info.find("bpm_max").text) / 100.0 # Grab valid difficulties for difficulty in music_entry.find("difficulty"): # Figure out the actual difficulty offset = { "novice": 0, "advanced": 1, "exhaust": 2, "infinite": 3, }.get(difficulty.tag) if offset is None: continue difficulties[offset] = int(difficulty.find("difnum").text) limited[offset] = int(difficulty.find("limited").text) # Import it self.start_batch() for chart in self.charts: # First, try to find in the DB from another version old_id = self.get_music_id_for_song(songid, chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {songid} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {songid} chart {chart}") next_id = old_id data = { "limited": limited[chart], "difficulty": difficulties[chart], "bpm_min": bpm_min, "bpm_max": bpm_max, } self.insert_music_id_for_song( next_id, songid, chart, title, artist, None, data ) self.finish_batch() def import_from_server(self, server: str, token: str) -> None: # First things first, lets try to import the music DB. We want to make # sure that even if the server doesn't respond right, we have a chart # entry for every chart for each song we're importing. music = self.remote_music(server, token) music_lut: Dict[int, Dict[int, Song]] = {} for entry in music.get_all_songs(self.game, self.version): if entry.id not in music_lut: music_lut[entry.id] = { chart: Song( entry.game, entry.version, entry.id, chart, entry.name, entry.artist, entry.genre, {}, ) for chart in self.charts } music_lut[entry.id][entry.chart] = entry # Import it for _, songs in music_lut.items(): self.start_batch() for _, song in songs.items(): # First, try to find in the DB from another version old_id = self.get_music_id_for_song(song.id, song.chart) if self.no_combine or old_id is None: # Insert original print(f"New entry for {song.id} chart {song.chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {song.id} chart {song.chart}") next_id = old_id data = { "limited": song.data.get_int("limited"), "difficulty": song.data.get_int("difficulty"), "bpm_min": song.data.get_int("bpm_min"), "bpm_max": song.data.get_int("bpm_max"), } self.insert_music_id_for_song( next_id, song.id, song.chart, song.name, song.artist, None, data ) self.finish_batch() class ReflecBeatScrapeConfiguration: def __init__( self, *, version: str, offset: int, stride: int, max_songs: int, max_difficulties: int, song_offset: int, song_length: int, artist_offset: Optional[int], artist_length: Optional[int], chart_offset: int, chart_length: int, difficulties_offset: int, ) -> None: self.version = version self.offset = offset self.stride = stride self.max_songs = max_songs self.max_difficulties = max_difficulties self.song_offset = song_offset self.song_length = song_length self.artist_offset = artist_offset self.artist_length = artist_length self.chart_offset = chart_offset self.chart_length = chart_length self.difficulties_offset = difficulties_offset class ImportReflecBeat(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: # We always have 4 charts, even if we're importing from Colette and below, # so that we guarantee a stable song ID. We'll be in trouble if Reflec # ever adds a fifth chart. if version in ["1", "2", "3", "4", "5", "6"]: actual_version = { "1": VersionConstants.REFLEC_BEAT, "2": VersionConstants.REFLEC_BEAT_LIMELIGHT, "3": VersionConstants.REFLEC_BEAT_COLETTE, "4": VersionConstants.REFLEC_BEAT_GROOVIN, "5": VersionConstants.REFLEC_BEAT_VOLZZA, "6": VersionConstants.REFLEC_BEAT_VOLZZA_2, }[version] self.charts = [0, 1, 2, 3] else: raise CLIException( "Unsupported ReflecBeat version, expected one of the following: 1, 2, 3, 4, 5, 6" ) super().__init__( config, GameConstants.REFLEC_BEAT, actual_version, no_combine, update ) def scrape(self, infile: str) -> List[Dict[str, Any]]: with open(infile, mode="rb") as myfile: data = myfile.read() myfile.close() configurations: List[ReflecBeatScrapeConfiguration] = [] if self.version == VersionConstants.REFLEC_BEAT: # Based on KBR:A:A:A:2011112300 configurations.append( ReflecBeatScrapeConfiguration( version="KBR:A:A:A:2011112300", offset=0xBFBD0, stride=280, max_songs=93, max_difficulties=3, song_offset=0x4C, song_length=0x40, # Artists aren't included in this mix. artist_offset=None, artist_length=None, chart_offset=0xD5, chart_length=0x20, difficulties_offset=0xD2, ) ) elif self.version == VersionConstants.REFLEC_BEAT_LIMELIGHT: # Based on LBR:A:A:A:2012082900 configurations.append( ReflecBeatScrapeConfiguration( version="LBR:A:A:A:2012082900", offset=0x132C48, stride=220, max_songs=191, max_difficulties=3, song_offset=0x4C, song_length=0x40, # Artists aren't included in this mix. artist_offset=None, artist_length=None, chart_offset=0x9B, chart_length=0x20, difficulties_offset=0x98, ) ) elif self.version == VersionConstants.REFLEC_BEAT_COLETTE: # Based on MBR:J:A:A:2014011600 configurations.append( ReflecBeatScrapeConfiguration( version="MBR:J:A:A:2014011600", offset=0x1E6880, stride=468, max_songs=443, max_difficulties=3, song_offset=0x34, song_length=0x80, artist_offset=0xB4, artist_length=0x80, chart_offset=0x1B4, chart_length=0x20, difficulties_offset=0x1A8, ) ) elif self.version == VersionConstants.REFLEC_BEAT_GROOVIN: # Based on MBR:J:A:A:2015102100 configurations.append( ReflecBeatScrapeConfiguration( version="MBR:J:A:A:2015102100", offset=0x212EC0, stride=524, max_songs=698, max_difficulties=4, song_offset=0x3C, song_length=0x80, artist_offset=0xBC, artist_length=0x80, chart_offset=0x1E8, chart_length=0x20, difficulties_offset=0x1D0, ) ) elif self.version == VersionConstants.REFLEC_BEAT_VOLZZA: # Based on MBR:J:A:A:2016030200 configurations.append( ReflecBeatScrapeConfiguration( version="MBR:J:A:A:2016030200", offset=0x1A0EC8, stride=552, max_songs=805, max_difficulties=4, song_offset=0x38, song_length=0x80, artist_offset=0xB8, artist_length=0x80, chart_offset=0x1E4, chart_length=0x20, difficulties_offset=0x1CC, ) ) elif self.version == VersionConstants.REFLEC_BEAT_VOLZZA_2: # Based on MBR:J:A:A:2016100400 configurations.append( ReflecBeatScrapeConfiguration( version="MBR:J:A:A:2016100400", offset=0x1CBC68, stride=552, max_songs=850, max_difficulties=4, song_offset=0x38, song_length=0x80, artist_offset=0xB8, artist_length=0x80, chart_offset=0x1E4, chart_length=0x20, difficulties_offset=0x1CC, ) ) else: raise CLIException(f"Unsupported ReflecBeat version {self.version}") def convert_string(inb: bytes) -> str: end = None for i in range(len(inb)): if inb[i] == 0: end = i break if end is None: raise UnicodeError("Invalid string!") if end == 0: return "" return inb[:end].decode("shift_jisx0213") def convert_version(songid: int, folder: int) -> int: if self.version == VersionConstants.REFLEC_BEAT_VOLZZA_2: # Reflec Volzza 2 appears from network and DLL perspective to be identical # to Volzza 1, including what version the game thinks it is for songs. So, # hard code the new song IDs so we can show the difference on the frontend. if folder == 5: if songid in [733, 760, 772, 773, 774, 782, 785, 786]: return 6 if songid >= 788: return 6 return folder for config in configurations: try: print(f"Trying configuration for game version {config.version}...") songs = [] for i in range(config.max_songs): start = config.offset + (config.stride * i) end = start + config.stride songdata = data[start:end] title = convert_string( songdata[ config.song_offset : ( config.song_offset + config.song_length ) ] ) if config.artist_offset is None: artist = "" else: artist = convert_string( songdata[ config.artist_offset : ( config.artist_offset + config.artist_length ) ] ) if title == "" and artist == "": continue songid = struct.unpack(" List[Dict[str, Any]]: # Grab music info from remote server music = self.remote_music(server, token) songs = music.get_all_songs(self.game, self.version) lut: Dict[int, Dict[str, Any]] = {} for song in songs: if song.id not in lut: lut[song.id] = { "id": song.id, "title": song.name, "artist": song.artist, "chartid": song.data.get_str("chart_id"), "difficulties": [0] * len(self.charts), "folder": song.data.get_int("folder"), } lut[song.id]["difficulties"][song.chart] = song.data.get_int("difficulty") # Return the reassembled data return [val for _, val in lut.items()] def import_music_db(self, songs: List[Dict[str, Any]]) -> None: for song in songs: self.start_batch() for chart in self.charts: songid = song["id"] chartid = song["chartid"] # ReflecBeat re-numbers some of their songs and overlaps with IDs from older # versions, so we need to keep a virtual mapping similar to DDR. Its not good # enough to just do title/artist because Reflec also has revival charts that # are named the same. Luckily we have internal chart ID to map on! old_id = self.get_music_id_for_song_data( None, None, chartid, chart, version=0 ) if self.no_combine or old_id is None: # Insert original print(f"New entry for {songid} chart {chart}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {songid} chart {chart}") next_id = old_id if old_id is None: # Add the virtual music entry we talked about above. Use the song ID when we discover # the song, plus the folder as a way to make them unique. self.insert_music_id_for_song( next_id, self.version * 10000 + songid, chart, song["title"], song["artist"], chartid, # Chart goes into genre for reflec, so we can handle revival charts { "difficulty": song["difficulties"][chart], "folder": song["folder"], "chart_id": chartid, }, version=0, ) else: if self.update: # Force a folder/difficulty update for this song. self.update_metadata_for_music_id( old_id, song["title"], song["artist"], chartid, # Chart goes into genre for reflec, so we can handle revival charts { "difficulty": song["difficulties"][chart], "folder": song["folder"], "chart_id": chartid, }, version=0, ) # Add the normal entry so the game finds the song. self.insert_music_id_for_song( next_id, songid, chart, song["title"], song["artist"], None, # Reflec Beat has no genres for real songs. { "difficulty": song["difficulties"][chart], "folder": song["folder"], "chart_id": chartid, }, ) self.finish_batch() class ImportDanceEvolution(ImportBase): def __init__( self, config: Config, version: str, no_combine: bool, update: bool, ) -> None: if version in ["1"]: actual_version = 1 else: raise CLIException( "Unsupported Dance Evolution version, expected one of the following: 1" ) super().__init__( config, GameConstants.DANCE_EVOLUTION, actual_version, no_combine, update ) def scrape(self, infile: str) -> List[Dict[str, Any]]: with open(infile, mode="rb") as myfile: data = myfile.read() myfile.close() arc = ARC(data) data = arc.read_file("data/song/song_params.plist") # First, do a header check like the game does if data[0:4] != b"MS02": raise Exception("Invalid song params file!") if data[4:6] not in [b"BE", b"LE"]: raise Exception("Invalid song params file!") def get_string(offset: int, default: Optional[str] = None) -> str: lut_offset = struct.unpack(">I", data[(offset) : (offset + 4)])[0] if lut_offset == 0: if default is None: raise Exception("Expecting a string, got empty!") return default length = 0 while data[lut_offset + length] != 0: length += 1 return ( data[lut_offset : (lut_offset + length)] .decode("utf-8") .replace("\n", " ") ) def get_int(offset: int) -> int: return struct.unpack(">I", data[(offset) : (offset + 4)])[0] # Now, make sure we know how long the file is numsongs = struct.unpack(">I", data[8:12])[0] filelen = struct.unpack(">I", data[12:16])[0] if filelen != len(data): raise Exception("Invalid song params file!") # Now, extract the meaningful data for each song retval = [] for i in range(numsongs): offset = (i * 128) + 16 songcode = get_string(offset + 0) # noqa: F841 songres1 = get_string(offset + 4) # noqa: F841 songres2 = get_string(offset + 8) # noqa: F841 bpm_min = get_int(offset + 12) bpm_max = get_int(offset + 16) copyright = get_string(offset + 24, "") title = get_string(offset + 52, "Unknown song") artist = get_string(offset + 56, "Unknown artist") level = get_int(offset + 64) charares1 = get_string(offset + 72) # noqa: F841 charares2 = get_string(offset + 76) # noqa: F841 kana_sort = get_string(offset + 108) flag1 = data[offset + 33] != 0x00 # noqa: F841 flag2 = data[offset + 34] == 0x01 # noqa: F841 flag3 = data[offset + 34] == 0x02 # noqa: F841 flag4 = data[offset + 116] != 0x00 # noqa: F841 # TODO: Get the real music ID from the data, once we have in-game traffic. retval.append( { "id": i, "title": title, "artist": artist, "copyright": copyright or None, "sort_key": kana_sort, "bpm_min": bpm_min, "bpm_max": bpm_max, "level": level, } ) return retval def lookup(self, server: str, token: str) -> List[Dict[str, Any]]: # TODO: We never got far enough to support DanEvo in the server, or # specify it in BEMAPI. So this is a dead function for now, but maybe # some year in the future I'll be able to support this. return [] def import_music_db(self, songs: List[Dict[str, Any]]) -> None: for song in songs: # Import it self.start_batch() # First, try to find in the DB from another version old_id = self.get_music_id_for_song(song["id"], 0) if self.no_combine or old_id is None: # Insert original print(f"New entry for {song['id']} chart {0}") next_id = self.get_next_music_id() else: # Insert pointing at same ID so scores transfer print(f"Reused entry for {song['id']} chart {0}") next_id = old_id data = { "level": song["level"], "bpm_min": song["bpm_min"], "bpm_max": song["bpm_max"], } self.insert_music_id_for_song( next_id, song["id"], 0, song["title"], song["artist"], None, data ) self.finish_batch() def main() -> None: parser = argparse.ArgumentParser(description="Import Game Music DB") parser.add_argument( "--series", action="store", type=str, required=True, help="The game series we are importing.", ) parser.add_argument( "--version", dest="version", action="store", type=str, required=True, help="The game version we are importing.", ) parser.add_argument( "--csv", dest="csv", action="store", type=str, help="The CSV file to read, for applicable games.", ) parser.add_argument( "--tsv", dest="tsv", action="store", type=str, help="The TSV file to read, for applicable games.", ) parser.add_argument( "--xml", dest="xml", action="store", type=str, help="The game XML file to read, for applicable games.", ) parser.add_argument( "--bin", dest="bin", action="store", type=str, help="The game binary file to read, for applicable games.", ) parser.add_argument( "--assets", dest="assets", action="store", type=str, help="The game sound assets directory, for applicable games.", ) parser.add_argument( "--no-combine", dest="no_combine", action="store_true", default=False, help="Don't look for the same music ID in other versions.", ) parser.add_argument( "--update", dest="update", action="store_true", default=False, help="Overwrite data with updated values when it already exists.", ) parser.add_argument( "--config", type=str, default="config.yaml", help="Core configuration for importing to DB. Defaults to 'config.yaml'.", ) parser.add_argument( "--server", type=str, help="The remote BEMAPI server to read data from.", ) parser.add_argument( "--token", type=str, help="The access token to use with the remote BEMAPI server.", ) parser.add_argument( "--folder", dest="folder", action="store", type=str, help="The path were a folder of files are stored.", ) # Parse args, validate invariants. args = parser.parse_args() if (args.token and not args.server) or (args.server and not args.token): raise CLIException("Must specify both --server and --token together!") if (args.csv or args.tsv or args.xml or args.bin or args.assets) and ( args.server or args.token ): raise CLIException( "Cannot specify both a remote server and a local file to read from!" ) # Load the config so we can talk to the server config = Config() load_config(args.config, config) series = None try: series = GameConstants(args.series) except ValueError: pass if series == GameConstants.POPN_MUSIC: popn = ImportPopn(config, args.version, args.no_combine, args.update) if args.bin: songs = popn.scrape(args.bin) if args.xml: songs = popn.scrape_xml(args.xml, songs) elif args.folder: files = Path(args.folder).glob("*xml") for filename in files: try: songs = popn.scrape_xml(filename, songs) except Exception: # We should really be just catching invalid XML but I didn't write this # nor do I have omnimix so I can't really test what it should do. raise CLIException(f"Invalid XML ({filename})") elif args.server and args.token: songs = popn.lookup(args.server, args.token) else: raise CLIException( "No game DLL provided and no remote server specified! Please " + "provide either a --bin or a --server and --token option!" ) popn.import_music_db(songs) popn.close() elif series == GameConstants.JUBEAT: jubeat = ImportJubeat(config, args.version, args.no_combine, args.update) if args.tsv is not None: # Special case for Jubeat, grab the title/artist metadata that was # hand-populated since its not in the music DB. jubeat.import_metadata(args.tsv) else: # Normal case, doing a music DB or emblem import. if args.xml is not None: songs, emblems = jubeat.scrape(args.xml) elif args.server and args.token: songs, emblems = jubeat.lookup(args.server, args.token) else: raise CLIException( "No music_info.xml or TSV provided and no remote server specified! Please " + "provide either a --xml, --tsv or a --server and --token option!" ) jubeat.import_music_db(songs) jubeat.import_emblems(emblems) jubeat.close() elif series == GameConstants.IIDX: iidx = ImportIIDX(config, args.version, args.no_combine, args.update) if args.tsv is not None: # Special case for IIDX, grab the title/artist metadata that was # wrong in the music DB, and correct it. iidx.import_metadata(args.tsv) else: # Normal case, doing a music DB import. if args.bin is not None: songs, qpros = iidx.scrape(args.bin, args.assets) elif args.server and args.token: songs, qpros = iidx.lookup(args.server, args.token) else: raise CLIException( "No music_data.bin or TSV provided and no remote server specified! Please " + "provide either a --bin, --tsv or a --server and --token option!" ) iidx.import_music_db(songs) iidx.import_qpros(qpros) iidx.close() elif series == GameConstants.DDR: ddr = ImportDDR(config, args.version, args.no_combine, args.update) if args.server and args.token: songs = ddr.lookup(args.server, args.token) else: if args.version == "16": if args.bin is None: raise CLIException("No startup.arc provided!") # DDR Ace has a different format altogether songs = ddr.parse_xml(args.bin) else: if args.bin is None: raise CLIException("No game DLL provided!") if args.xml is None: raise CLIException("No game music XML provided!") # DDR splits the music DB between the DLL and external XML # (Why??), so we must first scrape then hydrate with extra # data to get the full DB. songs = ddr.scrape(args.bin) songs = ddr.hydrate(songs, args.xml) ddr.import_music_db(songs) ddr.close() elif series == GameConstants.SDVX: sdvx = ImportSDVX(config, args.version, args.no_combine, args.update) if args.server and args.token: sdvx.import_from_server(args.server, args.token) else: if args.xml is None and args.bin is None and args.csv is None: raise CLIException( "No XML file or game DLL or appeal card CSV provided and " + "no remote server specified! Please provide either a --xml, " + "--bin, --csv or a --server and --token option!" ) if args.xml is not None: sdvx.import_music_db_or_appeal_cards(args.xml) if args.bin is not None: sdvx.import_catalog(args.bin) if args.csv is not None: sdvx.import_appeal_cards(args.csv) sdvx.close() elif series == GameConstants.MUSECA: museca = ImportMuseca(config, args.version, args.no_combine, args.update) if args.server and args.token: museca.import_from_server(args.server, args.token) elif args.xml is not None: museca.import_music_db(args.xml) else: raise CLIException( "No music-info.xml provided and no remote server specified! " + "Please provide either a --xml or a --server and --token option!" ) museca.close() elif series == GameConstants.REFLEC_BEAT: reflec = ImportReflecBeat(config, args.version, args.no_combine, args.update) if args.bin is not None: songs = reflec.scrape(args.bin) elif args.server and args.token: songs = reflec.lookup(args.server, args.token) else: raise CLIException( "No game DLL provided and no remote server specified! " + "Please provide either a --bin or a --server and --token option!" ) reflec.import_music_db(songs) reflec.close() elif series == GameConstants.DANCE_EVOLUTION: danevo = ImportDanceEvolution( config, args.version, args.no_combine, args.update ) if args.server and args.token: songs = danevo.lookup(args.server, args.token) elif args.bin is not None: songs = danevo.scrape(args.bin) else: raise CLIException( "No resource_lists.arc provided and no remote server " + "specified! Please provide either a --bin or a " + "--server and --token option!", ) danevo.import_music_db(songs) danevo.close() else: raise CLIException("Unsupported game series!") if __name__ == "__main__": try: main() except CLIException as e: # We terminate early with these when we want to show the user a message, but # not scare them with a full traceback. print(str(e))