# GNU MediaGoblin -- federated, autonomous media hosting
-# Copyright (C) 2011 MediaGoblin contributors. See AUTHORS.
+# Copyright (C) 2011, 2012 MediaGoblin contributors. See AUTHORS.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published by
import os
from mediagoblin import mg_globals as mgg
-
from mediagoblin.processing import BadMediaFail, \
create_pub_filepath, THUMB_SIZE, MEDIUM_SIZE
+from mediagoblin.tools.exif import exif_fix_image_orientation, \
+ extract_exif, clean_exif, get_gps_data, get_useful
+
+MAX_FILENAME_LENGTH = 255 # the limit in VFAT -- seems like a good baseline
+
+def resize_image(entry, filename, basename, file_tail, exif_tags, workdir,
+ new_size, size_limits=(0, 0)):
+ """Store a resized version of an image and return its pathname.
+
+ Arguments:
+ entry -- the entry for the image to resize
+ filename -- the filename of the original image being resized
+ basename -- simple basename of the given filename
+ file_tail -- ending string and extension for the resized filename
+ exif_tags -- EXIF data for the original image
+ workdir -- directory path for storing converted image files
+ new_size -- 2-tuple size for the resized image
+ size_limits (optional) -- image is only resized if it exceeds this size
-################################
-# Media processing initial steps
-################################
+ """
+ try:
+ resized = Image.open(filename)
+ except IOError:
+ raise BadMediaFail()
+ resized = exif_fix_image_orientation(resized, exif_tags) # Fix orientation
+ if ((resized.size[0] > size_limits[0]) or
+ (resized.size[1] > size_limits[1])):
+ resized.thumbnail(new_size, Image.ANTIALIAS)
+
+ # Truncate basename as needed so len(basename + file_tail) <= 255
+ resized_filename = (basename[:MAX_FILENAME_LENGTH - len(file_tail)] +
+ file_tail)
+ resized_filepath = create_pub_filepath(entry, resized_filename)
+
+ # Copy the new file to the conversion subdir, then remotely.
+ tmp_resized_filename = os.path.join(workdir, resized_filename)
+ with file(tmp_resized_filename, 'w') as resized_file:
+ resized.save(resized_file)
+ mgg.public_store.copy_local_to_storage(
+ tmp_resized_filename, resized_filepath)
+ return resized_filepath
def process_image(entry):
"""
workbench.dir, 'conversions')
os.mkdir(conversions_subdir)
- queued_filepath = entry['queued_media_file']
+ queued_filepath = entry.queued_media_file
queued_filename = workbench.localized_file(
mgg.queue_store, queued_filepath,
'source')
basename = os.path.split(filename_bits[0])[1]
extension = filename_bits[1].lower()
- try:
- thumb = Image.open(queued_filename)
- except IOError:
- raise BadMediaFail()
+ # EXIF extraction
+ exif_tags = extract_exif(queued_filename)
+ gps_data = get_gps_data(exif_tags)
- thumb.thumbnail(THUMB_SIZE, Image.ANTIALIAS)
-
- # Copy the thumb to the conversion subdir, then remotely.
- thumb_filename = 'thumbnail' + extension
- thumb_filepath = create_pub_filepath(entry, thumb_filename)
- tmp_thumb_filename = os.path.join(
- conversions_subdir, thumb_filename)
- with file(tmp_thumb_filename, 'w') as thumb_file:
- thumb.save(thumb_file)
- mgg.public_store.copy_local_to_storage(
- tmp_thumb_filename, thumb_filepath)
+ # Always create a small thumbnail
+ thumb_filepath = resize_image(entry, queued_filename, basename,
+ '.thumbnail' + extension, exif_tags,
+ conversions_subdir, THUMB_SIZE)
# If the size of the original file exceeds the specified size of a `medium`
- # file, a `medium.jpg` files is created and later associated with the media
+ # file, a `.medium.jpg` files is created and later associated with the media
# entry.
- medium = Image.open(queued_filename)
- medium_processed = False
-
- if medium.size[0] > MEDIUM_SIZE[0] or medium.size[1] > MEDIUM_SIZE[1]:
- medium.thumbnail(MEDIUM_SIZE, Image.ANTIALIAS)
-
- medium_filename = 'medium' + extension
- medium_filepath = create_pub_filepath(entry, medium_filename)
- tmp_medium_filename = os.path.join(
- conversions_subdir, medium_filename)
-
- with file(tmp_medium_filename, 'w') as medium_file:
- medium.save(medium_file)
-
- mgg.public_store.copy_local_to_storage(
- tmp_medium_filename, medium_filepath)
-
- medium_processed = True
+ medium_filepath = resize_image(entry, queued_filename, basename,
+ '.medium' + extension, exif_tags,
+ conversions_subdir, MEDIUM_SIZE, MEDIUM_SIZE)
# we have to re-read because unlike PIL, not everything reads
# things in string representation :)
as original_file:
original_file.write(queued_file.read())
+ # Remove queued media file from storage and database
mgg.queue_store.delete_file(queued_filepath)
- entry['queued_media_file'] = []
+ entry.queued_media_file = []
+
+ # Insert media file information into database
media_files_dict = entry.setdefault('media_files', {})
media_files_dict['thumb'] = thumb_filepath
media_files_dict['original'] = original_filepath
- if medium_processed:
- media_files_dict['medium'] = medium_filepath
+ media_files_dict['medium'] = medium_filepath
+
+ # Insert exif data into database
+ media_data = entry.setdefault('media_data', {})
+
+ # TODO: Fix for sql media_data, when exif is in sql
+ if media_data is not None:
+ media_data['exif'] = {
+ 'clean': clean_exif(exif_tags)}
+ media_data['exif']['useful'] = get_useful(
+ media_data['exif']['clean'])
+
+ if len(gps_data):
+ for key in list(gps_data.keys()):
+ gps_data['gps_' + key] = gps_data.pop(key)
+ entry.media_data_init(**gps_data)
# clean up workbench
workbench.destroy_self()
+
+if __name__ == '__main__':
+ import sys
+ import pprint
+
+ pp = pprint.PrettyPrinter()
+
+ result = extract_exif(sys.argv[1])
+ gps = get_gps_data(result)
+ clean = clean_exif(result)
+ useful = get_useful(clean)
+
+ print pp.pprint(
+ clean)