2016-08-17 18:33:10 +00:00
|
|
|
#!/usr/bin/python
|
|
|
|
|
|
|
|
import argparse
|
|
|
|
import datetime
|
|
|
|
import hashlib # md5
|
|
|
|
import lxml.etree as ET
|
|
|
|
import os
|
|
|
|
import re
|
|
|
|
import shutil
|
|
|
|
import subprocess
|
|
|
|
import time
|
2017-01-12 14:39:35 +00:00
|
|
|
import sgprops
|
2017-02-10 11:55:30 +00:00
|
|
|
import sys
|
2017-02-27 00:32:01 +00:00
|
|
|
import catalogTags
|
2016-08-17 18:33:10 +00:00
|
|
|
|
|
|
|
CATALOG_VERSION = 4
|
|
|
|
|
|
|
|
parser = argparse.ArgumentParser()
|
|
|
|
parser.add_argument("--update", help="Update/pull SCM source",
|
|
|
|
action="store_true")
|
|
|
|
parser.add_argument("--no-update",
|
|
|
|
help="Disable updating from SCM source",
|
|
|
|
action="store_true")
|
|
|
|
parser.add_argument("--clean", help="Force regeneration of all zip files",
|
|
|
|
action="store_true")
|
|
|
|
parser.add_argument("dir", help="Catalog directory")
|
|
|
|
args = parser.parse_args()
|
|
|
|
|
2017-01-12 14:39:35 +00:00
|
|
|
includes = []
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# xml node (robust) get text helper
|
|
|
|
def get_xml_text(e):
|
|
|
|
if e != None and e.text != None:
|
|
|
|
return e.text
|
|
|
|
else:
|
|
|
|
return ''
|
|
|
|
|
|
|
|
# create an xml node with text content
|
|
|
|
def make_xml_leaf(name, text):
|
|
|
|
leaf = ET.Element(name)
|
|
|
|
if text != None:
|
|
|
|
if isinstance(text, (int, long)):
|
|
|
|
leaf.text = str(text)
|
|
|
|
else:
|
|
|
|
leaf.text = text
|
|
|
|
else:
|
|
|
|
leaf.text = ''
|
|
|
|
return leaf
|
|
|
|
|
|
|
|
# return all available aircraft information from the set file as a
|
|
|
|
# dict
|
2017-01-12 14:39:35 +00:00
|
|
|
def scan_set_file(aircraft_dir, set_file):
|
|
|
|
global includes
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
base_file = os.path.basename(set_file)
|
|
|
|
base_id = base_file[:-8]
|
2017-01-12 14:39:35 +00:00
|
|
|
set_path = os.path.join(aircraft_dir, set_file)
|
|
|
|
|
|
|
|
local_includes = includes
|
|
|
|
local_includes.append(aircraft_dir)
|
|
|
|
root_node = sgprops.readProps(set_path, includePaths = local_includes)
|
|
|
|
|
|
|
|
if not root_node.hasChild("sim"):
|
|
|
|
return None
|
|
|
|
|
|
|
|
sim_node = root_node.getChild("sim")
|
2016-08-17 18:33:10 +00:00
|
|
|
if sim_node == None:
|
|
|
|
return None
|
2017-01-12 14:39:35 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
variant = {}
|
2017-01-12 14:39:35 +00:00
|
|
|
variant['name'] = sim_node.getValue("description", None)
|
|
|
|
variant['status'] = sim_node.getValue("status", None)
|
|
|
|
variant['author'] = sim_node.getValue("author", None)
|
|
|
|
variant['description'] = sim_node.getValue("long-description", None)
|
2016-08-17 18:33:10 +00:00
|
|
|
variant['id'] = base_id
|
2017-01-12 14:39:35 +00:00
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
# allow -set.xml files to declare themselves as primary.
|
|
|
|
# we use this avoid needing a variant-of in every other -set.xml
|
|
|
|
variant['primary-set'] = sim_node.getValue('primary-set', False)
|
|
|
|
|
|
|
|
# extract and record previews for each variant
|
|
|
|
if sim_node.hasChild('previews'):
|
2017-02-12 15:14:05 +00:00
|
|
|
print "has previews ..."
|
2017-01-12 13:42:08 +00:00
|
|
|
variant['previews'] = extract_previews(sim_node.getChild('previews'), aircraft_dir)
|
|
|
|
|
2017-01-12 14:39:35 +00:00
|
|
|
if sim_node.hasChild('rating'):
|
|
|
|
rating_node = sim_node.getChild("rating")
|
|
|
|
variant['rating_FDM'] = rating_node.getValue("FDM", 0)
|
|
|
|
variant['rating_systems'] = rating_node.getValue("systems", 0)
|
|
|
|
variant['rating_cockpit'] = rating_node.getValue("cockpit", 0)
|
|
|
|
variant['rating_model'] = rating_node.getValue("model", 0)
|
|
|
|
|
2017-02-27 00:32:01 +00:00
|
|
|
if sim_node.hasChild('tags'):
|
|
|
|
variant['tags'] = extract_tags(sim_node.getChild('tags'), set_file)
|
|
|
|
|
2017-01-12 14:39:35 +00:00
|
|
|
variant['variant-of'] = sim_node.getValue("variant-of", None)
|
2016-08-17 18:33:10 +00:00
|
|
|
#print ' ', variant
|
|
|
|
return variant
|
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
def extract_previews(previews_node, aircraft_dir):
|
|
|
|
result = []
|
|
|
|
for node in previews_node.getChildren("preview"):
|
|
|
|
previewType = node.getValue("type", None)
|
|
|
|
previewPath = node.getValue("path", None)
|
|
|
|
|
|
|
|
# check path exists in base-name-dir
|
|
|
|
fullPath = os.path.join(aircraft_dir, previewPath)
|
|
|
|
if not os.path.isfile(fullPath):
|
|
|
|
print "Bad preview path, skipping:" + fullPath
|
|
|
|
continue
|
|
|
|
result.append({'type':previewType, 'path':previewPath})
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
2017-02-27 00:32:01 +00:00
|
|
|
def extract_tags(tags_node, set_path):
|
|
|
|
result = []
|
|
|
|
for node in tags_node.getChildren("tag"):
|
|
|
|
tag = node.value
|
|
|
|
# check tag is in the allowed list
|
|
|
|
if not catalogTags.isValidTag(tag):
|
|
|
|
print "Unknown tag value:", tag, " in ", set_path
|
|
|
|
result.append(tag)
|
|
|
|
|
|
|
|
return result
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# scan all the -set.xml files in an aircraft directory. Returns a
|
|
|
|
# package dict and a list of variants.
|
|
|
|
def scan_aircraft_dir(aircraft_dir):
|
2017-01-12 13:42:08 +00:00
|
|
|
# old way of finding the master aircraft: it's the only one whose
|
|
|
|
# variant-of is empty. All the others have an actual value
|
|
|
|
# newer alternative is to specify one -set.xml as the primary. All the
|
|
|
|
# others are therefore variants.
|
|
|
|
setDicts = []
|
2016-08-17 18:33:10 +00:00
|
|
|
found_master = False
|
|
|
|
package = None
|
2017-01-12 13:42:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
files = os.listdir(aircraft_dir)
|
|
|
|
for file in sorted(files, key=lambda s: s.lower()):
|
|
|
|
if file.endswith('-set.xml'):
|
2017-01-12 14:39:35 +00:00
|
|
|
try:
|
2017-01-12 13:42:08 +00:00
|
|
|
d = scan_set_file(aircraft_dir, file)
|
|
|
|
if d == None:
|
2017-01-12 14:39:35 +00:00
|
|
|
continue
|
|
|
|
except:
|
2017-02-10 11:55:30 +00:00
|
|
|
print "Skipping set file since couldn't be parsed:", os.path.join(aircraft_dir, file), sys.exc_info()[0]
|
2017-01-12 13:42:08 +00:00
|
|
|
continue
|
2017-01-12 14:39:35 +00:00
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
setDicts.append(d)
|
|
|
|
if d['primary-set']:
|
|
|
|
found_master = True
|
|
|
|
package = d
|
|
|
|
|
|
|
|
# didn't find a dict identified explicitly as the primary, look for one
|
|
|
|
# with an undefined variant-of
|
|
|
|
if not found_master:
|
|
|
|
for d in setDicts:
|
|
|
|
if d['variant-of'] == '':
|
|
|
|
found_master = True
|
|
|
|
package = d
|
|
|
|
break
|
|
|
|
|
|
|
|
if not found_master:
|
|
|
|
if len(setDicts) > 1:
|
|
|
|
print "Warning, no explicit primary set.xml in " + aircraft_dir
|
|
|
|
# use the first one
|
|
|
|
package = setDicts[0]
|
|
|
|
|
|
|
|
# variants is just all the set dicts except the master
|
|
|
|
variants = setDicts
|
|
|
|
variants.remove(package)
|
2016-08-17 18:33:10 +00:00
|
|
|
return (package, variants)
|
|
|
|
|
|
|
|
# use svn commands to report the last change date within dir
|
|
|
|
def last_change_date_svn(dir):
|
|
|
|
command = [ 'svn', 'info', dir ]
|
|
|
|
result = subprocess.check_output( command )
|
|
|
|
match = re.search('Last Changed Date: (\d+)\-(\d+)\-(\d+)', result)
|
|
|
|
if match:
|
|
|
|
rev_str = match.group(1) + match.group(2) + match.group(3)
|
|
|
|
return int(rev_str)
|
|
|
|
|
|
|
|
# find the most recent mtime within a directory subtree
|
|
|
|
def scan_dir_for_change_date_mtime(path):
|
|
|
|
maxsec = 0
|
|
|
|
names = os.listdir(path)
|
|
|
|
for name in names:
|
|
|
|
fullname = os.path.join(path, name)
|
|
|
|
if name == '.' or name == '..':
|
|
|
|
pass
|
|
|
|
elif os.path.isdir( fullname ):
|
|
|
|
mtime = scan_dir_for_change_date_mtime( fullname )
|
|
|
|
if mtime > maxsec:
|
|
|
|
maxsec = mtime
|
|
|
|
else:
|
|
|
|
mtime = os.path.getmtime( fullname )
|
|
|
|
if mtime > maxsec:
|
|
|
|
maxsec = mtime
|
|
|
|
return maxsec
|
|
|
|
|
|
|
|
def make_aircraft_zip(repo_path, name, zip_file):
|
|
|
|
print "Updating:", name + '.zip'
|
2016-08-19 11:29:50 +00:00
|
|
|
savedir = os.getcwd()
|
2016-08-17 18:33:10 +00:00
|
|
|
os.chdir(repo_path)
|
|
|
|
if os.path.exists(zip_file):
|
|
|
|
os.remove(zip_file)
|
|
|
|
command = ['zip', '-rq', '-9']
|
|
|
|
if os.path.exists(zip_excludes):
|
|
|
|
command += ['-x@' + zip_excludes]
|
|
|
|
else:
|
|
|
|
print "warning: no zip-excludes.lst file provided", zip_excludes
|
|
|
|
command += [zip_file, name]
|
|
|
|
subprocess.call(command)
|
2016-08-19 11:29:50 +00:00
|
|
|
os.chdir(savedir)
|
2016-08-17 18:33:10 +00:00
|
|
|
|
|
|
|
def get_md5sum(file):
|
|
|
|
f = open(file, 'r')
|
|
|
|
md5sum = hashlib.md5(f.read()).hexdigest()
|
2016-08-19 11:29:50 +00:00
|
|
|
f.close()
|
2016-08-17 18:33:10 +00:00
|
|
|
return md5sum
|
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
def append_preview_nodes(node, variant, download_base, package_name):
|
|
|
|
if not 'previews' in variant:
|
|
|
|
return
|
|
|
|
|
|
|
|
for preview in variant['previews']:
|
|
|
|
preview_node = ET.Element('preview')
|
|
|
|
preview_url = download_base + 'previews/' + package_name + '_' + preview['path']
|
|
|
|
preview_node.append( make_xml_leaf('type', preview['type']) )
|
|
|
|
preview_node.append( make_xml_leaf('url', preview_url) )
|
|
|
|
preview_node.append( make_xml_leaf('path', preview['path']) )
|
|
|
|
node.append(preview_node)
|
|
|
|
|
|
|
|
def copy_previews_for_variant(variant, package_name, package_dir, previews_dir):
|
|
|
|
if not 'previews' in variant:
|
|
|
|
return
|
|
|
|
|
|
|
|
for preview in variant['previews']:
|
|
|
|
preview_src = os.path.join(package_dir, preview['path'])
|
|
|
|
preview_dst = os.path.join(previews_dir, package_name + '_' + preview['path'])
|
2017-02-12 15:14:05 +00:00
|
|
|
#print preview_src, preview_dst, preview['path']
|
|
|
|
dir = os.path.dirname(preview_dst)
|
|
|
|
if not os.path.isdir(dir):
|
|
|
|
os.makedirs(dir)
|
2017-01-12 13:42:08 +00:00
|
|
|
if os.path.exists(preview_src):
|
|
|
|
shutil.copy2(preview_src, preview_dst)
|
|
|
|
|
|
|
|
def copy_previews_for_package(package, variants, package_name, package_dir, previews_dir):
|
|
|
|
copy_previews_for_variant(package, package_name, package_dir, previews_dir)
|
|
|
|
for v in variants:
|
|
|
|
copy_previews_for_variant(v, package_name, package_dir, previews_dir)
|
|
|
|
|
2017-02-27 00:32:01 +00:00
|
|
|
def append_tag_nodes(node, variant):
|
|
|
|
if not 'tags' in variant:
|
|
|
|
return
|
|
|
|
|
|
|
|
for tag in variant['tags']:
|
|
|
|
node.append(make_xml_leaf('tag', tag))
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
#def get_file_stats(file):
|
|
|
|
# f = open(file, 'r')
|
|
|
|
# md5 = hashlib.md5(f.read()).hexdigest()
|
|
|
|
# file_size = os.path.getsize(file)
|
|
|
|
# return (md5, file_size)
|
|
|
|
|
|
|
|
if not os.path.isdir(args.dir):
|
|
|
|
print "A valid catalog directory must be provided"
|
|
|
|
exit(0)
|
|
|
|
|
|
|
|
parser = ET.XMLParser(remove_blank_text=True)
|
|
|
|
|
|
|
|
config_file = os.path.join(args.dir, 'catalog.config.xml')
|
|
|
|
config = ET.parse(config_file, parser)
|
|
|
|
config_node = config.getroot()
|
|
|
|
|
|
|
|
template_file = os.path.join(args.dir, 'template.xml')
|
|
|
|
template = ET.parse(template_file, parser)
|
|
|
|
template_root = template.getroot()
|
|
|
|
template_node = template_root.find('template')
|
|
|
|
|
|
|
|
md5sum_file = os.path.join(args.dir, 'md5sum.xml')
|
|
|
|
if os.path.exists(md5sum_file):
|
|
|
|
md5sum_tree = ET.parse(md5sum_file, parser)
|
|
|
|
md5sum_root = md5sum_tree.getroot()
|
|
|
|
else:
|
|
|
|
md5sum_root = ET.Element('PropertyList')
|
|
|
|
md5sum_tree = ET.ElementTree(md5sum_root)
|
|
|
|
|
|
|
|
scm_list = config_node.findall('scm')
|
|
|
|
upload_node = config_node.find('upload')
|
|
|
|
download_base = get_xml_text(config_node.find('download-url'))
|
|
|
|
output_dir = get_xml_text(config_node.find('local-output'))
|
|
|
|
if output_dir == '':
|
|
|
|
output_dir = os.path.join(args.dir, 'output')
|
|
|
|
if not os.path.isdir(output_dir):
|
|
|
|
os.mkdir(output_dir)
|
2017-01-12 13:42:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
thumbnail_dir = os.path.join(output_dir, 'thumbnails')
|
|
|
|
if not os.path.isdir(thumbnail_dir):
|
|
|
|
os.mkdir(thumbnail_dir)
|
2017-01-12 13:42:08 +00:00
|
|
|
|
|
|
|
previews_dir = os.path.join(output_dir, 'previews')
|
|
|
|
if not os.path.isdir(previews_dir):
|
|
|
|
os.mkdir(previews_dir)
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
tmp = os.path.join(args.dir, 'zip-excludes.lst')
|
|
|
|
zip_excludes = os.path.realpath(tmp)
|
|
|
|
|
2017-01-12 14:39:35 +00:00
|
|
|
for i in config_node.findall("include-dir"):
|
|
|
|
path = get_xml_text(i)
|
|
|
|
if not os.path.exists(path):
|
|
|
|
print "Skipping missing include path:", path
|
|
|
|
continue
|
|
|
|
includes.append(path)
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# freshen repositories
|
|
|
|
if args.no_update:
|
|
|
|
print 'Skipping repository updates.'
|
|
|
|
else:
|
|
|
|
cwd = os.getcwd()
|
|
|
|
for scm in scm_list:
|
|
|
|
repo_type = get_xml_text(scm.find('type'))
|
|
|
|
repo_path = get_xml_text(scm.find('path'))
|
2017-01-12 14:39:35 +00:00
|
|
|
includes.append(repo_path)
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
if repo_type == 'svn':
|
|
|
|
print 'SVN update:', repo_path
|
|
|
|
subprocess.call(['svn', 'update', repo_path])
|
|
|
|
elif repo_type == 'git':
|
|
|
|
print 'GIT pull:', repo_path
|
|
|
|
os.chdir(repo_path)
|
|
|
|
subprocess.call(['git','pull'])
|
|
|
|
elif repo_type == 'no-scm':
|
|
|
|
print "No update of unmannaged files:", repo_path
|
|
|
|
else:
|
|
|
|
print "Unknown scm type:", scm, repo_path
|
|
|
|
os.chdir(cwd)
|
|
|
|
|
|
|
|
# names of zip files we want (so we can identify/remove orphans)
|
|
|
|
valid_zips = []
|
|
|
|
|
|
|
|
# create the catalog tree
|
|
|
|
catalog_node = ET.Element('PropertyList')
|
|
|
|
catalog_root = ET.ElementTree(catalog_node)
|
|
|
|
|
|
|
|
# include the template configuration
|
|
|
|
for child in template_node:
|
|
|
|
catalog_node.append(child)
|
|
|
|
|
|
|
|
# scan repositories for catalog information
|
|
|
|
for scm in scm_list:
|
|
|
|
repo_type = get_xml_text(scm.find('type'))
|
|
|
|
repo_path = get_xml_text(scm.find('path'))
|
|
|
|
skip_nodes = scm.findall('skip')
|
|
|
|
skip_list = []
|
|
|
|
for s in skip_nodes:
|
|
|
|
skip_list.append(get_xml_text(s))
|
|
|
|
print 'skip list:', skip_list
|
|
|
|
names = os.listdir(repo_path)
|
|
|
|
for name in sorted(names, key=lambda s: s.lower()):
|
|
|
|
if name in skip_list:
|
|
|
|
print "skipping:", name
|
|
|
|
continue
|
2017-01-12 13:42:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
aircraft_dir = os.path.join(repo_path, name)
|
|
|
|
if os.path.isdir(aircraft_dir):
|
|
|
|
print "%s:" % name,
|
|
|
|
(package, variants) = scan_aircraft_dir(aircraft_dir)
|
|
|
|
if package == None:
|
|
|
|
print "skipping:", name, "(no -set.xml files)"
|
|
|
|
continue
|
|
|
|
#print "package:", package
|
|
|
|
#print "variants:", variants
|
|
|
|
package_node = ET.Element('package')
|
|
|
|
package_node.append( make_xml_leaf('name', package['name']) )
|
|
|
|
package_node.append( make_xml_leaf('status', package['status']) )
|
|
|
|
package_node.append( make_xml_leaf('author', package['author']) )
|
|
|
|
package_node.append( make_xml_leaf('description', package['description']) )
|
|
|
|
if 'rating_FDM' in package or 'rating_systems' in package \
|
|
|
|
or 'rating_cockpit' in package or 'rating_model' in package:
|
|
|
|
rating_node = ET.Element('rating')
|
|
|
|
package_node.append(rating_node)
|
|
|
|
rating_node.append( make_xml_leaf('FDM',
|
|
|
|
package['rating_FDM']) )
|
|
|
|
rating_node.append( make_xml_leaf('systems',
|
|
|
|
package['rating_systems']) )
|
|
|
|
rating_node.append( make_xml_leaf('cockpit',
|
|
|
|
package['rating_cockpit']) )
|
|
|
|
rating_node.append( make_xml_leaf('model',
|
|
|
|
package['rating_model']) )
|
|
|
|
package_node.append( make_xml_leaf('id', package['id']) )
|
|
|
|
for variant in variants:
|
|
|
|
variant_node = ET.Element('variant')
|
|
|
|
package_node.append(variant_node)
|
|
|
|
variant_node.append( make_xml_leaf('id', variant['id']) )
|
|
|
|
variant_node.append( make_xml_leaf('name', variant['name']) )
|
2016-11-12 15:00:08 +00:00
|
|
|
if 'description' in variant:
|
|
|
|
variant_node.append( make_xml_leaf('description', variant['description']) )
|
|
|
|
|
2016-11-25 21:52:42 +00:00
|
|
|
if 'author' in variant:
|
2016-11-12 15:00:08 +00:00
|
|
|
variant_node.append( make_xml_leaf('author', variant['author']) )
|
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
append_preview_nodes(variant_node, variant, download_base, name)
|
2017-02-27 00:32:01 +00:00
|
|
|
append_tag_nodes(variant_node, variant)
|
2017-01-12 13:42:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
package_node.append( make_xml_leaf('dir', name) )
|
|
|
|
if not download_base.endswith('/'):
|
|
|
|
download_base += '/'
|
|
|
|
download_url = download_base + name + '.zip'
|
|
|
|
thumbnail_url = download_base + 'thumbnails/' + name + '_thumbnail.jpg'
|
|
|
|
package_node.append( make_xml_leaf('url', download_url) )
|
|
|
|
package_node.append( make_xml_leaf('thumbnail', thumbnail_url) )
|
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
append_preview_nodes(package_node, package, download_base, name)
|
2017-02-27 00:32:01 +00:00
|
|
|
append_tag_nodes(package_node, package)
|
2017-01-12 13:42:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# todo: url (download), thumbnail (download url)
|
|
|
|
|
|
|
|
# get cached md5sum if it exists
|
|
|
|
md5sum = get_xml_text(md5sum_root.find(str('aircraft_' + name)))
|
|
|
|
|
|
|
|
# now do the packaging and rev number stuff
|
|
|
|
dir_mtime = scan_dir_for_change_date_mtime(aircraft_dir)
|
|
|
|
if repo_type == 'svn':
|
|
|
|
rev = last_change_date_svn(aircraft_dir)
|
|
|
|
else:
|
|
|
|
d = datetime.datetime.utcfromtimestamp(dir_mtime)
|
|
|
|
rev = d.strftime("%Y%m%d")
|
|
|
|
package_node.append( make_xml_leaf('revision', rev) )
|
|
|
|
#print "rev:", rev
|
|
|
|
#print "dir mtime:", dir_mtime
|
|
|
|
zipfile = os.path.join( output_dir, name + '.zip' )
|
|
|
|
valid_zips.append(name + '.zip')
|
|
|
|
if not os.path.exists(zipfile) \
|
|
|
|
or dir_mtime > os.path.getmtime(zipfile) \
|
|
|
|
or args.clean:
|
|
|
|
# rebuild zip file
|
|
|
|
print "updating:", zipfile
|
|
|
|
make_aircraft_zip(repo_path, name, zipfile)
|
|
|
|
md5sum = get_md5sum(zipfile)
|
|
|
|
else:
|
|
|
|
print "(no change)"
|
|
|
|
if md5sum == "":
|
|
|
|
md5sum = get_md5sum(zipfile)
|
|
|
|
filesize = os.path.getsize(zipfile)
|
|
|
|
package_node.append( make_xml_leaf('md5', md5sum) )
|
|
|
|
package_node.append( make_xml_leaf('file-size-bytes', filesize) )
|
|
|
|
|
|
|
|
# handle md5sum cache
|
|
|
|
node = md5sum_root.find('aircraft_' + name)
|
|
|
|
if node != None:
|
|
|
|
node.text = md5sum
|
|
|
|
else:
|
|
|
|
md5sum_root.append( make_xml_leaf('aircraft_' + name, md5sum) )
|
2016-11-12 15:00:08 +00:00
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# handle thumbnails
|
|
|
|
thumbnail_src = os.path.join(aircraft_dir, 'thumbnail.jpg')
|
|
|
|
thumbnail_dst = os.path.join(thumbnail_dir, name + '_thumbnail.jpg')
|
|
|
|
if os.path.exists(thumbnail_src):
|
|
|
|
shutil.copy2(thumbnail_src, thumbnail_dst)
|
|
|
|
catalog_node.append(package_node)
|
|
|
|
package_node.append( make_xml_leaf('thumbnail-path', 'thumbnail.jpg') )
|
|
|
|
|
2017-01-12 13:42:08 +00:00
|
|
|
# copy previews for the package and variants into the
|
|
|
|
# output directory
|
|
|
|
copy_previews_for_package(package, variants, name, aircraft_dir, previews_dir)
|
|
|
|
|
2016-08-17 18:33:10 +00:00
|
|
|
# write out the master catalog file
|
|
|
|
cat_file = os.path.join(output_dir, 'catalog.xml')
|
|
|
|
catalog_root.write(cat_file, encoding='utf-8', xml_declaration=True, pretty_print=True)
|
|
|
|
|
|
|
|
# write out the md5sum cache file
|
|
|
|
print md5sum_file
|
|
|
|
md5sum_tree.write(md5sum_file, encoding='utf-8', xml_declaration=True, pretty_print=True)
|
|
|
|
|
|
|
|
# look for orphaned zip files
|
|
|
|
files = os.listdir(output_dir)
|
|
|
|
for file in files:
|
|
|
|
if file.endswith('.zip')and not file in valid_zips:
|
|
|
|
print "orphaned zip:", file
|