SlideShare ist ein Scribd-Unternehmen logo
1 von 4
Downloaden Sie, um offline zu lesen
R2P: Recomposition and Retargeting of
Photographic Images
Hui-Tang Chang
Po-Cheng Pan
EE, National Taiwan
University, Taipei, Taiwan
r02921089@ntu.edu.tw
b00901087@ntu.edu.tw
Yu-Chiang Frank Wang
Research Center for IT
Innovation, Academia Sinica,
Taipei, Taiwan
ycwang@citi.sinica.edu.tw
Ming-Syan Chen
Electrical Engineering,
National Taiwan University,
Taipei, Taiwan
mschen@cc.ee.ntu.edu.tw
ABSTRACT
In this paper, we propose a novel approach for perform-
ing joint recomposition and retargeting of photographic im-
ages (R2P). Given a reference image of interest, our method
is able to automatically alter the composition of the input
source image accordingly, while the recomposed output will
be jointly retargeted to fit the reference. This is achieved
by recomposing the visual components of the source image
via graph matching, followed by solving a constrained mesh-
warping based optimization problem for retargeting. As a
result, the recomposed output image would fit the refer-
ence while suppressing possible distortion. Our experiments
confirm that our proposed R2P method is able to achieve
visually satisfactory results, without the need to use pre-
collected labeled data or predetermined aesthetics rules.
Categories and Subject Descriptors
I.4.3 [Image Processing & computer vision]: Enhance-
ment; I.4.9 [Image Processing & computer vision]: Ap-
plications; I.5.4 [Pattern Recognition]: Applications -
Computer Vision
General Terms
Algorithms, Experimentation
Keywords
Photography Composition; Graph Matching; Image Aes-
thetics; Computational Photography
1. INTRODUCTION
Image enhancement aims at improving the visual qual-
ity of an image, which can be achieved by applying photo-
graphic techniques (e.g., high dynamic range (HDR) or auto
white balance (AWB)) or image processing algorithms (e.g.,
Permission to make digital or hard copies of all or part of this work for personal or
classroom use is granted without fee provided that copies are not made or distributed
for profit or commercial advantage and that copies bear this notice and the full cita-
tion on the first page. Copyrights for components of this work owned by others than
ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or re-
publish, to post on servers or to redistribute to lists, requires prior specific permission
and/or a fee. Request permissions from Permissions@acm.org.
MM’15, October 26-30, 2015, Brisbane, Australia
c 2015 ACM. ISBN 978-1-4503-3459-4/15/10 ...$15.00.
DOI: http://dx.doi.org/10.1145/2733373.2806366.
Reference ImageSource Image Output Image
R2P Process
Figure 1: Example of a source image which is recomposed
and retargeted to fit the reference image.
denoising or deblurring). However the only thing which can-
not be easily altered is the geometric structure of visual com-
ponents, i.e., composition, of an image.
Improving image aesthetics is an alternative way to in-
crease user satisfaction. Several computer vision and ma-
chine learning techniques have been applied to address this
task. Since composition of an image has been viewed as a
factor in assessing the image aesthetics, some researchers ad-
dressed the task of photographic composition [10, 9, 1, 6, 7].
Most of them aim at cropping the input image for satisfying
particular rules of image aesthetics. For example, Nishiyama
et al. [7] proposed sensation-based cropping for learning im-
age aesthetics models. Santella et al. [8] utilized eye-tracking
data, while Liu et al. [6] used pre-defined photography com-
position rules for image cropping. In addition to image crop-
ping, Wong and Wong [9] proposed an algorithm to adjust
the ratio of foreground/background regions. To perform re-
composition, Bhattacharya et al. [1] trained their aesthet-
ics models using pre-collected labeled (scored) image data,
while Zhang et al. [10] further took the relationship between
visual components into consideration.
Nevertheless, the above methods require pre-determined
rules or pre-collected training data for performing photog-
raphy recomposition. In practice, one cannot apply such
rules to images taken in different scenarios. Thus, the ro-
bustness of such methods would not be sufficient. Recently,
Chang et al. [2] proposed a transfer-learning based method
for recompose a source image based on a reference image of
interest. However, their approach applied image inpainting
techniques which are computationally expensive. Moreover,
they need to fine-tune their output image by post-processing
the recomposed foreground regions.
Inspired by a mesh-based image retargetting work [5], we
extend the work of [2] by integrating image retargeting and
warping into a unified framework for image recomposition.
As illustrated in Figure 1, we not only transfer the photo-
927
(a) Mesh Generation (b) Salience Map (c) Object Extraction (d) Initial Mesh (e) Recomposed Mesh
Figure 2: Illustration of foreground extraction and recomposition. Note that M(u) and ˆM(u) denote the source and recom-
posed locations of the uth mesh point in the background, respectively.
graphic composition from the reference to the source image,
the output image will be simultaneously retargeted to fit the
reference. Thus, we refer to our proposed methods as recom-
position and retargeting of photographic images (R2P).
In Section 2, we will present the details of our proposed
method for joint recomposition and retargeting. We will
discuss that, while no user interaction or pre-collected la-
beled training data are required, why our method is able to
preserve image context information after the R2P process
without additional post-processing procedures. It is worth
repeating that, our work focus on transferring the photogra-
phy composition from a reference image to the source image,
rather than improving image aesthetic scores for the input
image by pre-determined rules or features.
The contributions of our work are highlighted below:
• Automatically transfer the photography composition
from a reference to the source image by solving a graph-
matching based optimization problem.
• With the observed composition information, we per-
form a joint recomposition and retargeting on the source
image, so that it would visually fit the reference image.
• Our proposed method is fully automatic, without the
need to use pre-collected labeled training data, pre-
determine aesthetics rules, or user interaction.
2. OUR PROPOSED FRAMEWORK
2.1 Foreground Object Retrieval
Given a source and reference image pair, we first identify
the foreground objects in both images. Prior works like [2]
and [10] applied saliency detection and graph-cut for iter-
atively detect the foreground regions. For simplicity, we
consider an one-pass algorithm as described below.
Inspired by a mesh-based image retargeting algorithm [5],
we apply Canny edge detectors to detect the edges in an im-
age, followed by sampling the pixels along the detected edges
as the feature points. We note that, as depicted in Figure 2a,
we mark the features points of detected edges and bound-
aries in green and blue, respectively. In order to describe
smooth regions (e.g., sky in Figure 2a), we additionally add
random feature points for collecting a sufficient number for
representation purposes. Once such feature points are ob-
tained, we utilize Delaunay triangulation algorithm to con-
struct triangular meshes, as shown in Figure 2a.
Instead of iteratively applying saliency detection for de-
termining possible foreground objects, our method directly
combines the derived meshes and the saliency detection out-
put of [4] (e.g., Figure 2b) for separating foreground and
background regions, which is achieved by DB-scan cluster-
ing. Finally, N foreground objects set can be automatically
observed. Also, we record all strong edges detected in the
background regions (see Figure 2c), which will be crucial for
the final stage of joint recomposition and retargeting.
2.2 Recomposition via Graph Matching
In our work, we apply a composition graph G = (V, E, A)
to describe each foreground object, and the recomposition
can be performed by graph matching. For each graph, V
and E denote the foreground objects and the edges con-
necting between them, respectively. We have A contain the
attributes as for each vertex and al for each edge. That is,
for Vx ∈ V , as(x) indicates the area of that region. For each
edge exy ∈ E, we have al(x, y) measuring the normalized
Euclidean distance between the two vertices of edge exy.
To perform photography recomposition, we solve a graph-
matching based optimization problem based on the observed
affinity matrix W between the two composition graphs GS
and GR
, which are the constructed by source and reference
images, respectively. The affinity matrix W is composed of
the affinity between vertices and edges in the dimension of
|V S
||V R
| × |V S
||V R
|. In W, each entry w is defined as:
wxm;yn =



f(vS
x , vR
m) if x = y, m = n
f(eS
xy, eR
mn) if x = y, m = n
0 otherwise,
(1)
where the affinity functions are calculated as follows:
f(vS
x , vR
m) = min[as(x), as(m)]/max[as(x), as(m)],
f(eS
xy, eR
mn) = 1/ aI
l (x, y) − aR
l (m, n) .
(2)
Now, we convert the original recomposition problem into
the task of determining an indicator matrix Z ∈ {0, 1}|V S
|×|V R
|
,
where zxm = 1 indicates a matching of foreground regions
between V S
x and V R
m . For computation purposes, we make
Z into a column vector z ∈ {0, 1}|V S
||V R
|
, and apply the
method of [3] for solve the following problem instead:
max zT
Wz , s.t.
|V R
|
k=1
zxk ≤ 1,
|V I
|
k=1
zkm ≤ 1, ∀x, m.
(3)
The constraints in (3) enforces one-to-one matching between
GS
to GR
. After solving (3), the locations of the foreground
objects in the source image can be updated accordingly. As
depicted in Figure 3, we denote the original location of ob-
jects in the source image as S = [s1, s2, ...sN ] ∈ R2×N
, the
estimated ones as P = [p1, p2, ...pN ] ∈ R2×N
, and the fi-
nal recomposed and retargeted ones as Q = [q1, q2, ...qN ] ∈
R2×N
(as detailed in the following subsection).
928
Source Image Reference Image
s1
Recomposed Image
s2
s3
s4
p1
p2 p4
p3
q1
q2 q4
q3
Figure 3: Graph matching for recomposition. The blue lines
denote the composition graphs, and the red dotted lines as-
sociate the foreground objects across images. Note that s,
p, and q denote the object locations in each image.
x
A31
A12
A23
M(u3
)
M(u1
)
M(u2
)
A31 A12
A23
M(u1
)
x
^
M(u3
)
^ M(u2
)
^
Figure 4: Illustration of mesh-based texture mapping.
2.3 Joint Recomposition and Retargeting
2.3.1 Background Mesh Warping
From Section 2.2, we estimate the new locations of the
foreground objects in the source image for recomposition
purposes. To complete the proposed R2P process, we pro-
pose to advance background warping for joint recomposition
and retargeting. This allows us to move the foreground ob-
jects to the target locations without leaving any unfilled or
distorted regions in the background.
We note that, the recent works of [2] and [10] apply im-
age inpainting techniques for producing their recomposed
outputs. Such techniques require the background regions to
exhibit proper textural properties. Moreover, it is also com-
putationally expensive to perform image inpainting (com-
pared to our warping-based algorithm).
In [5], a uniform stretch method was considered to con-
struct image meshes for image retargeting. Their goal was
to minimize the predicted error of the recovered image with
consistency in the observed boundaries and edges. Since
their approach is not able to alter the composition of the
input image (i.e., recomposition), we propose a novel mesh-
warping algorithm for solving the tasks of recomposition and
retargeting simultaneously.
To be more specific, our goal is to estimate the locations
of foreground objects and the image cropping factor, which
would that fulfill the following two requirements:
1. Recompose and retarget the input to fit the reference.
2. Minimize the distortion of the output image.
With the above two goals, our task is to determine the op-
timal locations Q for each foreground object and the image
cropping factor b, which are defined as follows:
Q = [q1, q2, ...qN ] ∈ R2×N
b = [btop, bbottom, bleft, bright] ∈ R1×4
,
(4)
where each b indicates the amount of the image boundary
to be cropped (negative b values mean image stretching).
Thus, the objective function to be optimized for joint re-
composition and retargeting is:
[Q, b] = argmin
Q,b
D(Q, b)E(Q, b). (5)
Algorithm 1 Optimization for Background Mesh Warping
1: Sort O in decreasing order of object size.
2: for all u ∈ M, n ∈ [1, N] do
3: cn(u) = 1/distance(u, On)2.
4: end for
5: for all Possible decision of b do
6: for n = 1 to N do
7: for all qn, qn − pn < rmax do
8:
ˆMn(u) =
cn(u) M(u) + qn − sn +
k=1→n−1
ck(u) ˆMk(u)
k=1→n
ck(u)
, ∀u
9: Handle boundary constraint.
10: Calculate D(qn, b) and E(qn, b).
11: end for
12: ˆqn = argminqn D(qn, b)E(qn, b)
13: end for
14: Qb = [ˆq1, ˆq2, ...ˆqN ]
15: end for
16: Find ˆB = argminb D(Qb, b)E(Qb, b)
17: return ˆQ = Q(qˆb, ˆb)
The first objective function D jointly solve recomposition
and retargeting as follows:
D(Q, b) =
n∈[1,N]
(1 + α qn − pn )
bi∈b
(1 + β bi ), (6)
To avoid remarkable changes between the original and out-
put images, we set the constraint of the foreground object
locations in the output image as qn − pn < rmax. We
also limit the cropping factor bi < bmax. In our work,
parameters rmax and bmax are set as 10% and 20% of the
width of the source image, respectively.
On the other hand, the objective function E suppresses
image distortion, which is determined by:
E(Q, b) = (1 + γF)
Lx∈L (u,v)∈Lx
ˆM(u, v) − M(u, v)
2
,
(7)
where M and ˆM represent the 2D coordinates of source and
recomposed mesh points (see Figure 2d and 2e for exam-
ples), and u and v denote the indices of the corresponding
mesh points. Thus, M(u, v) represents the vector connect-
ing points u and v. F is the proportion of flip-over triangle
in the recomposed output, L denotes the set of strong edges
(see Section 2.1).
To solve (5) for producing the final output image, we cal-
culate the distances between each mesh point and the fore-
ground objects, and update the locations of the mesh points
ˆM (via iterating between the calculation of q and b). Note
that, for display purposes, we show the recomposed but un-
cropped output in Figure 2e (i.e., no updates of b). As for
satisfying the boundary constraints to minimize image dis-
tortion, we uniformly stretch the background mesh grids to
meet the nearest image boundary. The optimization process
is summarized in Algorithm 1.
2.3.2 Texture Mapping for Image Completion
Once the locations of each mesh point in ˆM are deter-
mined, the remaining task is to render the final image out-
put. This is achieved by texture mapping. That is, for each
929
(a)
(b)
(c)
(d)
(e)
Source Image
Output Image Reference Image
Figure 5: Examples of photography recomposition and re-
targeting with different foreground object numbers.
triangle in M, we calculate the areas A12, A23 and A31 for
each pixel x inside the triangular mesh (see Figure 4 for il-
lustration). By barycentric interpolation, the pixels in ˆM
can be computed as follows:
ˆx = A23
ˆM(u1) + A31
ˆM(u2) + A12
ˆM(u3) /(A23+A31+A12).
(8)
After all pixels are derived, the output will be the final re-
composed and retargeted image.
3. EXPERIMENTS
We consider a variety of photographic images for eval-
uation. All the image data are collected from either those
considered in related works or the Internet1
. Figure 5 shows
the results produced by our method. It can be seen that,
given different reference images, our method successfully re-
composed and retargeted images with single or multiple fore-
ground objects. For example, we were able to transfer the
composition and preserve the foreground-to-background re-
lationships (see the shadow of the cat in Figure 5b and the
grass regions in Figure 5c). As noted in [2], we can only
provide qualitative evaluation for the effectiveness of recom-
position, since no aesthetic scores can be calculated (we do
not apply any predetermined aesthetic rules).
Compared to a recent work of [2], they required complex
post-processing procedures to refine the recomposed outputs
to avoid possible background mismatch. Moreover, they did
not address the task of retargeting, so their outputs cannot
fit the references well (see Figure 6 for example).
Finally, we present results in Figure 7 to show that our
method can retarget the images, so that the outputs would
1
Fig.1-s: CC Attribution, Thomas Leuthard; Fig.5a-r: hjjanisch @Flickr;
Fig.5b-s, 5d-s: cited from [10]; Others: free or CC0 or public image from In-
ternet.
Results of [2]. Our output image
Figure 6: Comparison between [2] and our approach. Note
that [2] simply performs image inpainting for recomposition,
and thus its output does not fit the reference in Figure 5c.
Figure 7: Retargeted outputs of Figure 5d using reference
images (in black and white) with different aspect ratios.
fit the reference images with different aspect ratios. From
the above experiments, we confirm the the effectiveness and
robustness of our proposed method for joint recomposition
and retargeting photographic images.
4. CONCLUSION
We presented an automatic framework to perform joint
recomposition and retargeting of the input image for fitting
the reference of interest. Unlike prior works requiring pre-
collected data or predetermined aesthetics rules, our method
applies graph matching to predict the locations of the fore-
ground objects, while both foreground and background re-
gions will be retargeted to fit the reference image. The
above process is completed by solving a constrained mesh-
warping based optimization problem, which minimizes both
recovered errors and image distortion. Our experiments on
several real-world photographic images with different fore-
ground numbers verified the effectiveness of our method.
Acknowledgement This work is supported in part by
the Ministry of Science and Technology of Taiwan via MOST104-
3011-f-002-005 and MOST103-2221-E-001-021-MY2.
5. REFERENCE
[1] S. Bhattacharya, R. Sukthankar, and M. Shah. A framework
for photo-quality assessment and enhancement based on visual
aesthetics. In ACM Multimedia, 2010.
[2] H.-T. Chang, Y.-C. F. Wang, and M.-S. Chen. Transfer in pho-
tography composition. In ACM Multimedia, 2014.
[3] T. Cour et al. Balanced graph matching. In NIPS. 2007.
[4] S. Goferman, L. Zelnik-Manor, and A. Tal. Context-aware
saliency detection. In IEEE CVPR, 2010.
[5] Y. Guo et al. Image retargeting using mesh parametrization.
IEEE Trans. Multimedia, 2009.
[6] L. Liu, R. Chen, L. Wolf, and D. Cohen-Or. Optimizing photo
composition. Computer Graphics Forum, 2010.
[7] M. Nishiyama, T. Okabe, Y. Sato, and I. Sato. Sensation-based
photo cropping. In ACM Multimedia, 2009.
[8] A. Santella et al. Gaze-based interaction for semi-automatic
photo cropping. In SIGCHI, 2006.
[9] L.-K. Wong and K.-L. Wong. Enhancing visual dominance by
semantics-preserving image recomposition. In ACM Multimedia,
2012.
[10] F.-L. Zhang, M. Wang, and S.-M. Hu. Aesthetic image enhance-
ment by dependence-aware object recomposition. IEEE Trans.
Multimedia, 2013.
930

Weitere ähnliche Inhalte

Was ist angesagt?

G143741
G143741G143741
G143741irjes
 
Intel ILS: Enhancing Photorealism Enhancement
Intel ILS: Enhancing Photorealism EnhancementIntel ILS: Enhancing Photorealism Enhancement
Intel ILS: Enhancing Photorealism EnhancementAlejandro Franceschi
 
Tracking Faces using Active Appearance Models
Tracking Faces using Active Appearance ModelsTracking Faces using Active Appearance Models
Tracking Faces using Active Appearance ModelsComponica LLC
 
Gil Shapira's Active Appearance Model slides
Gil Shapira's Active Appearance Model slidesGil Shapira's Active Appearance Model slides
Gil Shapira's Active Appearance Model slideswolf
 
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...Alejandro Franceschi
 
Structure from motion
Structure from motionStructure from motion
Structure from motionFatima Radi
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD Editor
 
Ijarcet vol-2-issue-7-2319-2322
Ijarcet vol-2-issue-7-2319-2322Ijarcet vol-2-issue-7-2319-2322
Ijarcet vol-2-issue-7-2319-2322Editor IJARCET
 
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONING
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONINGFRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONING
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONINGcscpconf
 
Fractal image compression with adaptive quardtree partitioning
Fractal image compression with adaptive quardtree partitioningFractal image compression with adaptive quardtree partitioning
Fractal image compression with adaptive quardtree partitioningcsandit
 
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...Habibur Rahman
 
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...Editor IJCATR
 
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...Image Processing using Matlab ( using a built in Matlab function(Histogram eq...
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...Majd Khaleel
 
Use of horizontal and vertical edge processing technique to improve number pl...
Use of horizontal and vertical edge processing technique to improve number pl...Use of horizontal and vertical edge processing technique to improve number pl...
Use of horizontal and vertical edge processing technique to improve number pl...eSAT Journals
 
Paper id 2420148
Paper id 2420148Paper id 2420148
Paper id 2420148IJRAT
 
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...CSCJournals
 
Image segmentation ppt
Image segmentation pptImage segmentation ppt
Image segmentation pptGichelle Amon
 
Image parts and segmentation
Image parts and segmentation Image parts and segmentation
Image parts and segmentation Rappy Saha
 
A version of watershed algorithm for color image segmentation
A version of watershed algorithm for color image segmentationA version of watershed algorithm for color image segmentation
A version of watershed algorithm for color image segmentationHabibur Rahman
 

Was ist angesagt? (20)

G143741
G143741G143741
G143741
 
Intel ILS: Enhancing Photorealism Enhancement
Intel ILS: Enhancing Photorealism EnhancementIntel ILS: Enhancing Photorealism Enhancement
Intel ILS: Enhancing Photorealism Enhancement
 
Tracking Faces using Active Appearance Models
Tracking Faces using Active Appearance ModelsTracking Faces using Active Appearance Models
Tracking Faces using Active Appearance Models
 
Gil Shapira's Active Appearance Model slides
Gil Shapira's Active Appearance Model slidesGil Shapira's Active Appearance Model slides
Gil Shapira's Active Appearance Model slides
 
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...
Google Research Siggraph Whitepaper | Total Relighting: Learning to Relight P...
 
Structure from motion
Structure from motionStructure from motion
Structure from motion
 
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
IJERD (www.ijerd.com) International Journal of Engineering Research and Devel...
 
Ijarcet vol-2-issue-7-2319-2322
Ijarcet vol-2-issue-7-2319-2322Ijarcet vol-2-issue-7-2319-2322
Ijarcet vol-2-issue-7-2319-2322
 
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONING
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONINGFRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONING
FRACTAL IMAGE COMPRESSION WITH ADAPTIVE QUARDTREE PARTITIONING
 
Fractal image compression with adaptive quardtree partitioning
Fractal image compression with adaptive quardtree partitioningFractal image compression with adaptive quardtree partitioning
Fractal image compression with adaptive quardtree partitioning
 
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...
Segmentation of Color Image using Adaptive Thresholding and Masking with Wate...
 
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...
An Improved Image Fusion Scheme Based on Markov Random Fields with Image Enha...
 
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...Image Processing using Matlab ( using a built in Matlab function(Histogram eq...
Image Processing using Matlab ( using a built in Matlab function(Histogram eq...
 
Use of horizontal and vertical edge processing technique to improve number pl...
Use of horizontal and vertical edge processing technique to improve number pl...Use of horizontal and vertical edge processing technique to improve number pl...
Use of horizontal and vertical edge processing technique to improve number pl...
 
Dv34745751
Dv34745751Dv34745751
Dv34745751
 
Paper id 2420148
Paper id 2420148Paper id 2420148
Paper id 2420148
 
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...
A Novel Super Resolution Algorithm Using Interpolation and LWT Based Denoisin...
 
Image segmentation ppt
Image segmentation pptImage segmentation ppt
Image segmentation ppt
 
Image parts and segmentation
Image parts and segmentation Image parts and segmentation
Image parts and segmentation
 
A version of watershed algorithm for color image segmentation
A version of watershed algorithm for color image segmentationA version of watershed algorithm for color image segmentation
A version of watershed algorithm for color image segmentation
 

Andere mochten auch

Supply chain planner kpi
Supply chain planner kpiSupply chain planner kpi
Supply chain planner kpijonhsfer
 
Top 8 supply chain consultant resume samples
Top 8 supply chain consultant resume samplesTop 8 supply chain consultant resume samples
Top 8 supply chain consultant resume samplestonychoper1305
 
William Menke Sr Operations Manager Resume
William Menke Sr Operations Manager ResumeWilliam Menke Sr Operations Manager Resume
William Menke Sr Operations Manager ResumeBill Menke
 
J. Keith Hubbard Resume - Manufacturing and Supply Chain Leadership
J. Keith Hubbard Resume -  Manufacturing and Supply Chain LeadershipJ. Keith Hubbard Resume -  Manufacturing and Supply Chain Leadership
J. Keith Hubbard Resume - Manufacturing and Supply Chain LeadershipJ. Keith Hubbard
 
Resume - Steve K. Mu - Supply Chain Engineer 0611
Resume - Steve K. Mu - Supply Chain Engineer 0611Resume - Steve K. Mu - Supply Chain Engineer 0611
Resume - Steve K. Mu - Supply Chain Engineer 0611Ko-Ping Mu
 
Analyst - Supply Chain Management
Analyst - Supply Chain ManagementAnalyst - Supply Chain Management
Analyst - Supply Chain ManagementChirag Dabhee
 
Supply Chain Manager
Supply Chain Manager Supply Chain Manager
Supply Chain Manager Gary Rice
 
Jerry James Supply Chain Manager Resume
Jerry James Supply Chain Manager ResumeJerry James Supply Chain Manager Resume
Jerry James Supply Chain Manager Resumejamescpim
 
Top 12 supply chain resume tips
Top 12 supply chain resume tipsTop 12 supply chain resume tips
Top 12 supply chain resume tipssupplychaincareer
 
Supply Chain Management, Procurement, Sourcing, Acquire, and Logistics
Supply Chain Management, Procurement, Sourcing, Acquire, and LogisticsSupply Chain Management, Procurement, Sourcing, Acquire, and Logistics
Supply Chain Management, Procurement, Sourcing, Acquire, and LogisticsRishabh Agarwal
 

Andere mochten auch (13)

Supply chain planner kpi
Supply chain planner kpiSupply chain planner kpi
Supply chain planner kpi
 
Top 8 supply chain consultant resume samples
Top 8 supply chain consultant resume samplesTop 8 supply chain consultant resume samples
Top 8 supply chain consultant resume samples
 
Suchismita saha
Suchismita sahaSuchismita saha
Suchismita saha
 
Executive Profile
Executive ProfileExecutive Profile
Executive Profile
 
William Menke Sr Operations Manager Resume
William Menke Sr Operations Manager ResumeWilliam Menke Sr Operations Manager Resume
William Menke Sr Operations Manager Resume
 
scott planner resume - Copy
scott planner resume - Copyscott planner resume - Copy
scott planner resume - Copy
 
J. Keith Hubbard Resume - Manufacturing and Supply Chain Leadership
J. Keith Hubbard Resume -  Manufacturing and Supply Chain LeadershipJ. Keith Hubbard Resume -  Manufacturing and Supply Chain Leadership
J. Keith Hubbard Resume - Manufacturing and Supply Chain Leadership
 
Resume - Steve K. Mu - Supply Chain Engineer 0611
Resume - Steve K. Mu - Supply Chain Engineer 0611Resume - Steve K. Mu - Supply Chain Engineer 0611
Resume - Steve K. Mu - Supply Chain Engineer 0611
 
Analyst - Supply Chain Management
Analyst - Supply Chain ManagementAnalyst - Supply Chain Management
Analyst - Supply Chain Management
 
Supply Chain Manager
Supply Chain Manager Supply Chain Manager
Supply Chain Manager
 
Jerry James Supply Chain Manager Resume
Jerry James Supply Chain Manager ResumeJerry James Supply Chain Manager Resume
Jerry James Supply Chain Manager Resume
 
Top 12 supply chain resume tips
Top 12 supply chain resume tipsTop 12 supply chain resume tips
Top 12 supply chain resume tips
 
Supply Chain Management, Procurement, Sourcing, Acquire, and Logistics
Supply Chain Management, Procurement, Sourcing, Acquire, and LogisticsSupply Chain Management, Procurement, Sourcing, Acquire, and Logistics
Supply Chain Management, Procurement, Sourcing, Acquire, and Logistics
 

Ähnlich wie p927-chang

Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...
Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...
Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...IJCSIS Research Publications
 
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an ObjectAnkur Tyagi
 
Local Binary Fitting Segmentation by Cooperative Quantum Particle Optimization
Local Binary Fitting Segmentation by Cooperative Quantum Particle OptimizationLocal Binary Fitting Segmentation by Cooperative Quantum Particle Optimization
Local Binary Fitting Segmentation by Cooperative Quantum Particle OptimizationTELKOMNIKA JOURNAL
 
Qcce quality constrained co saliency estimation for common object detection
Qcce quality constrained co saliency estimation for common object detectionQcce quality constrained co saliency estimation for common object detection
Qcce quality constrained co saliency estimation for common object detectionKoteswar Rao Jerripothula
 
Asilomar09 compressive superres
Asilomar09 compressive superresAsilomar09 compressive superres
Asilomar09 compressive superresHoàng Sơn
 
Structure and Motion - 3D Reconstruction of Cameras and Structure
Structure and Motion - 3D Reconstruction of Cameras and StructureStructure and Motion - 3D Reconstruction of Cameras and Structure
Structure and Motion - 3D Reconstruction of Cameras and StructureGiovanni Murru
 
An Hypergraph Object Oriented Model For Image Segmentation And Annotation
An Hypergraph Object Oriented Model For Image Segmentation And AnnotationAn Hypergraph Object Oriented Model For Image Segmentation And Annotation
An Hypergraph Object Oriented Model For Image Segmentation And AnnotationCrystal Sanchez
 
Image recogonization
Image recogonizationImage recogonization
Image recogonizationSANTOSH RATH
 
A Survey on Image Retrieval By Different Features and Techniques
A Survey on Image Retrieval By Different Features and TechniquesA Survey on Image Retrieval By Different Features and Techniques
A Survey on Image Retrieval By Different Features and TechniquesIRJET Journal
 
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKING
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKINGA PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKING
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKINGIRJET Journal
 
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion Network
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion NetworkIRJET- Optimization of Semantic Image Retargeting by using Guided Fusion Network
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion NetworkIRJET Journal
 
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...cscpconf
 
Semantic Image Retrieval Using Relevance Feedback
Semantic Image Retrieval Using Relevance Feedback  Semantic Image Retrieval Using Relevance Feedback
Semantic Image Retrieval Using Relevance Feedback dannyijwest
 
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient Boosting
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient BoostingDetection of Seam Carving in Uncompressed Images using eXtreme Gradient Boosting
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient BoostingIJCSIS Research Publications
 

Ähnlich wie p927-chang (20)

Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...
Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...
Image Stitching Algorithm: An Optimization between Correlation-Based and Feat...
 
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
 
1288
12881288
1288
 
Final
FinalFinal
Final
 
Real-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU AccelerationReal-Time Face Tracking with GPU Acceleration
Real-Time Face Tracking with GPU Acceleration
 
CUDA Accelerated Face Recognition
CUDA Accelerated Face RecognitionCUDA Accelerated Face Recognition
CUDA Accelerated Face Recognition
 
Local Binary Fitting Segmentation by Cooperative Quantum Particle Optimization
Local Binary Fitting Segmentation by Cooperative Quantum Particle OptimizationLocal Binary Fitting Segmentation by Cooperative Quantum Particle Optimization
Local Binary Fitting Segmentation by Cooperative Quantum Particle Optimization
 
Qcce quality constrained co saliency estimation for common object detection
Qcce quality constrained co saliency estimation for common object detectionQcce quality constrained co saliency estimation for common object detection
Qcce quality constrained co saliency estimation for common object detection
 
Asilomar09 compressive superres
Asilomar09 compressive superresAsilomar09 compressive superres
Asilomar09 compressive superres
 
06466595
0646659506466595
06466595
 
Structure and Motion - 3D Reconstruction of Cameras and Structure
Structure and Motion - 3D Reconstruction of Cameras and StructureStructure and Motion - 3D Reconstruction of Cameras and Structure
Structure and Motion - 3D Reconstruction of Cameras and Structure
 
An Hypergraph Object Oriented Model For Image Segmentation And Annotation
An Hypergraph Object Oriented Model For Image Segmentation And AnnotationAn Hypergraph Object Oriented Model For Image Segmentation And Annotation
An Hypergraph Object Oriented Model For Image Segmentation And Annotation
 
Seminarpaper
SeminarpaperSeminarpaper
Seminarpaper
 
Image recogonization
Image recogonizationImage recogonization
Image recogonization
 
A Survey on Image Retrieval By Different Features and Techniques
A Survey on Image Retrieval By Different Features and TechniquesA Survey on Image Retrieval By Different Features and Techniques
A Survey on Image Retrieval By Different Features and Techniques
 
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKING
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKINGA PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKING
A PROJECT REPORT ON REMOVAL OF UNNECESSARY OBJECTS FROM PHOTOS USING MASKING
 
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion Network
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion NetworkIRJET- Optimization of Semantic Image Retargeting by using Guided Fusion Network
IRJET- Optimization of Semantic Image Retargeting by using Guided Fusion Network
 
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...
EXTENDED WAVELET TRANSFORM BASED IMAGE INPAINTING ALGORITHM FOR NATURAL SCENE...
 
Semantic Image Retrieval Using Relevance Feedback
Semantic Image Retrieval Using Relevance Feedback  Semantic Image Retrieval Using Relevance Feedback
Semantic Image Retrieval Using Relevance Feedback
 
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient Boosting
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient BoostingDetection of Seam Carving in Uncompressed Images using eXtreme Gradient Boosting
Detection of Seam Carving in Uncompressed Images using eXtreme Gradient Boosting
 

p927-chang

  • 1. R2P: Recomposition and Retargeting of Photographic Images Hui-Tang Chang Po-Cheng Pan EE, National Taiwan University, Taipei, Taiwan r02921089@ntu.edu.tw b00901087@ntu.edu.tw Yu-Chiang Frank Wang Research Center for IT Innovation, Academia Sinica, Taipei, Taiwan ycwang@citi.sinica.edu.tw Ming-Syan Chen Electrical Engineering, National Taiwan University, Taipei, Taiwan mschen@cc.ee.ntu.edu.tw ABSTRACT In this paper, we propose a novel approach for perform- ing joint recomposition and retargeting of photographic im- ages (R2P). Given a reference image of interest, our method is able to automatically alter the composition of the input source image accordingly, while the recomposed output will be jointly retargeted to fit the reference. This is achieved by recomposing the visual components of the source image via graph matching, followed by solving a constrained mesh- warping based optimization problem for retargeting. As a result, the recomposed output image would fit the refer- ence while suppressing possible distortion. Our experiments confirm that our proposed R2P method is able to achieve visually satisfactory results, without the need to use pre- collected labeled data or predetermined aesthetics rules. Categories and Subject Descriptors I.4.3 [Image Processing & computer vision]: Enhance- ment; I.4.9 [Image Processing & computer vision]: Ap- plications; I.5.4 [Pattern Recognition]: Applications - Computer Vision General Terms Algorithms, Experimentation Keywords Photography Composition; Graph Matching; Image Aes- thetics; Computational Photography 1. INTRODUCTION Image enhancement aims at improving the visual qual- ity of an image, which can be achieved by applying photo- graphic techniques (e.g., high dynamic range (HDR) or auto white balance (AWB)) or image processing algorithms (e.g., Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full cita- tion on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or re- publish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from Permissions@acm.org. MM’15, October 26-30, 2015, Brisbane, Australia c 2015 ACM. ISBN 978-1-4503-3459-4/15/10 ...$15.00. DOI: http://dx.doi.org/10.1145/2733373.2806366. Reference ImageSource Image Output Image R2P Process Figure 1: Example of a source image which is recomposed and retargeted to fit the reference image. denoising or deblurring). However the only thing which can- not be easily altered is the geometric structure of visual com- ponents, i.e., composition, of an image. Improving image aesthetics is an alternative way to in- crease user satisfaction. Several computer vision and ma- chine learning techniques have been applied to address this task. Since composition of an image has been viewed as a factor in assessing the image aesthetics, some researchers ad- dressed the task of photographic composition [10, 9, 1, 6, 7]. Most of them aim at cropping the input image for satisfying particular rules of image aesthetics. For example, Nishiyama et al. [7] proposed sensation-based cropping for learning im- age aesthetics models. Santella et al. [8] utilized eye-tracking data, while Liu et al. [6] used pre-defined photography com- position rules for image cropping. In addition to image crop- ping, Wong and Wong [9] proposed an algorithm to adjust the ratio of foreground/background regions. To perform re- composition, Bhattacharya et al. [1] trained their aesthet- ics models using pre-collected labeled (scored) image data, while Zhang et al. [10] further took the relationship between visual components into consideration. Nevertheless, the above methods require pre-determined rules or pre-collected training data for performing photog- raphy recomposition. In practice, one cannot apply such rules to images taken in different scenarios. Thus, the ro- bustness of such methods would not be sufficient. Recently, Chang et al. [2] proposed a transfer-learning based method for recompose a source image based on a reference image of interest. However, their approach applied image inpainting techniques which are computationally expensive. Moreover, they need to fine-tune their output image by post-processing the recomposed foreground regions. Inspired by a mesh-based image retargetting work [5], we extend the work of [2] by integrating image retargeting and warping into a unified framework for image recomposition. As illustrated in Figure 1, we not only transfer the photo- 927
  • 2. (a) Mesh Generation (b) Salience Map (c) Object Extraction (d) Initial Mesh (e) Recomposed Mesh Figure 2: Illustration of foreground extraction and recomposition. Note that M(u) and ˆM(u) denote the source and recom- posed locations of the uth mesh point in the background, respectively. graphic composition from the reference to the source image, the output image will be simultaneously retargeted to fit the reference. Thus, we refer to our proposed methods as recom- position and retargeting of photographic images (R2P). In Section 2, we will present the details of our proposed method for joint recomposition and retargeting. We will discuss that, while no user interaction or pre-collected la- beled training data are required, why our method is able to preserve image context information after the R2P process without additional post-processing procedures. It is worth repeating that, our work focus on transferring the photogra- phy composition from a reference image to the source image, rather than improving image aesthetic scores for the input image by pre-determined rules or features. The contributions of our work are highlighted below: • Automatically transfer the photography composition from a reference to the source image by solving a graph- matching based optimization problem. • With the observed composition information, we per- form a joint recomposition and retargeting on the source image, so that it would visually fit the reference image. • Our proposed method is fully automatic, without the need to use pre-collected labeled training data, pre- determine aesthetics rules, or user interaction. 2. OUR PROPOSED FRAMEWORK 2.1 Foreground Object Retrieval Given a source and reference image pair, we first identify the foreground objects in both images. Prior works like [2] and [10] applied saliency detection and graph-cut for iter- atively detect the foreground regions. For simplicity, we consider an one-pass algorithm as described below. Inspired by a mesh-based image retargeting algorithm [5], we apply Canny edge detectors to detect the edges in an im- age, followed by sampling the pixels along the detected edges as the feature points. We note that, as depicted in Figure 2a, we mark the features points of detected edges and bound- aries in green and blue, respectively. In order to describe smooth regions (e.g., sky in Figure 2a), we additionally add random feature points for collecting a sufficient number for representation purposes. Once such feature points are ob- tained, we utilize Delaunay triangulation algorithm to con- struct triangular meshes, as shown in Figure 2a. Instead of iteratively applying saliency detection for de- termining possible foreground objects, our method directly combines the derived meshes and the saliency detection out- put of [4] (e.g., Figure 2b) for separating foreground and background regions, which is achieved by DB-scan cluster- ing. Finally, N foreground objects set can be automatically observed. Also, we record all strong edges detected in the background regions (see Figure 2c), which will be crucial for the final stage of joint recomposition and retargeting. 2.2 Recomposition via Graph Matching In our work, we apply a composition graph G = (V, E, A) to describe each foreground object, and the recomposition can be performed by graph matching. For each graph, V and E denote the foreground objects and the edges con- necting between them, respectively. We have A contain the attributes as for each vertex and al for each edge. That is, for Vx ∈ V , as(x) indicates the area of that region. For each edge exy ∈ E, we have al(x, y) measuring the normalized Euclidean distance between the two vertices of edge exy. To perform photography recomposition, we solve a graph- matching based optimization problem based on the observed affinity matrix W between the two composition graphs GS and GR , which are the constructed by source and reference images, respectively. The affinity matrix W is composed of the affinity between vertices and edges in the dimension of |V S ||V R | × |V S ||V R |. In W, each entry w is defined as: wxm;yn =    f(vS x , vR m) if x = y, m = n f(eS xy, eR mn) if x = y, m = n 0 otherwise, (1) where the affinity functions are calculated as follows: f(vS x , vR m) = min[as(x), as(m)]/max[as(x), as(m)], f(eS xy, eR mn) = 1/ aI l (x, y) − aR l (m, n) . (2) Now, we convert the original recomposition problem into the task of determining an indicator matrix Z ∈ {0, 1}|V S |×|V R | , where zxm = 1 indicates a matching of foreground regions between V S x and V R m . For computation purposes, we make Z into a column vector z ∈ {0, 1}|V S ||V R | , and apply the method of [3] for solve the following problem instead: max zT Wz , s.t. |V R | k=1 zxk ≤ 1, |V I | k=1 zkm ≤ 1, ∀x, m. (3) The constraints in (3) enforces one-to-one matching between GS to GR . After solving (3), the locations of the foreground objects in the source image can be updated accordingly. As depicted in Figure 3, we denote the original location of ob- jects in the source image as S = [s1, s2, ...sN ] ∈ R2×N , the estimated ones as P = [p1, p2, ...pN ] ∈ R2×N , and the fi- nal recomposed and retargeted ones as Q = [q1, q2, ...qN ] ∈ R2×N (as detailed in the following subsection). 928
  • 3. Source Image Reference Image s1 Recomposed Image s2 s3 s4 p1 p2 p4 p3 q1 q2 q4 q3 Figure 3: Graph matching for recomposition. The blue lines denote the composition graphs, and the red dotted lines as- sociate the foreground objects across images. Note that s, p, and q denote the object locations in each image. x A31 A12 A23 M(u3 ) M(u1 ) M(u2 ) A31 A12 A23 M(u1 ) x ^ M(u3 ) ^ M(u2 ) ^ Figure 4: Illustration of mesh-based texture mapping. 2.3 Joint Recomposition and Retargeting 2.3.1 Background Mesh Warping From Section 2.2, we estimate the new locations of the foreground objects in the source image for recomposition purposes. To complete the proposed R2P process, we pro- pose to advance background warping for joint recomposition and retargeting. This allows us to move the foreground ob- jects to the target locations without leaving any unfilled or distorted regions in the background. We note that, the recent works of [2] and [10] apply im- age inpainting techniques for producing their recomposed outputs. Such techniques require the background regions to exhibit proper textural properties. Moreover, it is also com- putationally expensive to perform image inpainting (com- pared to our warping-based algorithm). In [5], a uniform stretch method was considered to con- struct image meshes for image retargeting. Their goal was to minimize the predicted error of the recovered image with consistency in the observed boundaries and edges. Since their approach is not able to alter the composition of the input image (i.e., recomposition), we propose a novel mesh- warping algorithm for solving the tasks of recomposition and retargeting simultaneously. To be more specific, our goal is to estimate the locations of foreground objects and the image cropping factor, which would that fulfill the following two requirements: 1. Recompose and retarget the input to fit the reference. 2. Minimize the distortion of the output image. With the above two goals, our task is to determine the op- timal locations Q for each foreground object and the image cropping factor b, which are defined as follows: Q = [q1, q2, ...qN ] ∈ R2×N b = [btop, bbottom, bleft, bright] ∈ R1×4 , (4) where each b indicates the amount of the image boundary to be cropped (negative b values mean image stretching). Thus, the objective function to be optimized for joint re- composition and retargeting is: [Q, b] = argmin Q,b D(Q, b)E(Q, b). (5) Algorithm 1 Optimization for Background Mesh Warping 1: Sort O in decreasing order of object size. 2: for all u ∈ M, n ∈ [1, N] do 3: cn(u) = 1/distance(u, On)2. 4: end for 5: for all Possible decision of b do 6: for n = 1 to N do 7: for all qn, qn − pn < rmax do 8: ˆMn(u) = cn(u) M(u) + qn − sn + k=1→n−1 ck(u) ˆMk(u) k=1→n ck(u) , ∀u 9: Handle boundary constraint. 10: Calculate D(qn, b) and E(qn, b). 11: end for 12: ˆqn = argminqn D(qn, b)E(qn, b) 13: end for 14: Qb = [ˆq1, ˆq2, ...ˆqN ] 15: end for 16: Find ˆB = argminb D(Qb, b)E(Qb, b) 17: return ˆQ = Q(qˆb, ˆb) The first objective function D jointly solve recomposition and retargeting as follows: D(Q, b) = n∈[1,N] (1 + α qn − pn ) bi∈b (1 + β bi ), (6) To avoid remarkable changes between the original and out- put images, we set the constraint of the foreground object locations in the output image as qn − pn < rmax. We also limit the cropping factor bi < bmax. In our work, parameters rmax and bmax are set as 10% and 20% of the width of the source image, respectively. On the other hand, the objective function E suppresses image distortion, which is determined by: E(Q, b) = (1 + γF) Lx∈L (u,v)∈Lx ˆM(u, v) − M(u, v) 2 , (7) where M and ˆM represent the 2D coordinates of source and recomposed mesh points (see Figure 2d and 2e for exam- ples), and u and v denote the indices of the corresponding mesh points. Thus, M(u, v) represents the vector connect- ing points u and v. F is the proportion of flip-over triangle in the recomposed output, L denotes the set of strong edges (see Section 2.1). To solve (5) for producing the final output image, we cal- culate the distances between each mesh point and the fore- ground objects, and update the locations of the mesh points ˆM (via iterating between the calculation of q and b). Note that, for display purposes, we show the recomposed but un- cropped output in Figure 2e (i.e., no updates of b). As for satisfying the boundary constraints to minimize image dis- tortion, we uniformly stretch the background mesh grids to meet the nearest image boundary. The optimization process is summarized in Algorithm 1. 2.3.2 Texture Mapping for Image Completion Once the locations of each mesh point in ˆM are deter- mined, the remaining task is to render the final image out- put. This is achieved by texture mapping. That is, for each 929
  • 4. (a) (b) (c) (d) (e) Source Image Output Image Reference Image Figure 5: Examples of photography recomposition and re- targeting with different foreground object numbers. triangle in M, we calculate the areas A12, A23 and A31 for each pixel x inside the triangular mesh (see Figure 4 for il- lustration). By barycentric interpolation, the pixels in ˆM can be computed as follows: ˆx = A23 ˆM(u1) + A31 ˆM(u2) + A12 ˆM(u3) /(A23+A31+A12). (8) After all pixels are derived, the output will be the final re- composed and retargeted image. 3. EXPERIMENTS We consider a variety of photographic images for eval- uation. All the image data are collected from either those considered in related works or the Internet1 . Figure 5 shows the results produced by our method. It can be seen that, given different reference images, our method successfully re- composed and retargeted images with single or multiple fore- ground objects. For example, we were able to transfer the composition and preserve the foreground-to-background re- lationships (see the shadow of the cat in Figure 5b and the grass regions in Figure 5c). As noted in [2], we can only provide qualitative evaluation for the effectiveness of recom- position, since no aesthetic scores can be calculated (we do not apply any predetermined aesthetic rules). Compared to a recent work of [2], they required complex post-processing procedures to refine the recomposed outputs to avoid possible background mismatch. Moreover, they did not address the task of retargeting, so their outputs cannot fit the references well (see Figure 6 for example). Finally, we present results in Figure 7 to show that our method can retarget the images, so that the outputs would 1 Fig.1-s: CC Attribution, Thomas Leuthard; Fig.5a-r: hjjanisch @Flickr; Fig.5b-s, 5d-s: cited from [10]; Others: free or CC0 or public image from In- ternet. Results of [2]. Our output image Figure 6: Comparison between [2] and our approach. Note that [2] simply performs image inpainting for recomposition, and thus its output does not fit the reference in Figure 5c. Figure 7: Retargeted outputs of Figure 5d using reference images (in black and white) with different aspect ratios. fit the reference images with different aspect ratios. From the above experiments, we confirm the the effectiveness and robustness of our proposed method for joint recomposition and retargeting photographic images. 4. CONCLUSION We presented an automatic framework to perform joint recomposition and retargeting of the input image for fitting the reference of interest. Unlike prior works requiring pre- collected data or predetermined aesthetics rules, our method applies graph matching to predict the locations of the fore- ground objects, while both foreground and background re- gions will be retargeted to fit the reference image. The above process is completed by solving a constrained mesh- warping based optimization problem, which minimizes both recovered errors and image distortion. Our experiments on several real-world photographic images with different fore- ground numbers verified the effectiveness of our method. Acknowledgement This work is supported in part by the Ministry of Science and Technology of Taiwan via MOST104- 3011-f-002-005 and MOST103-2221-E-001-021-MY2. 5. REFERENCE [1] S. Bhattacharya, R. Sukthankar, and M. Shah. A framework for photo-quality assessment and enhancement based on visual aesthetics. In ACM Multimedia, 2010. [2] H.-T. Chang, Y.-C. F. Wang, and M.-S. Chen. Transfer in pho- tography composition. In ACM Multimedia, 2014. [3] T. Cour et al. Balanced graph matching. In NIPS. 2007. [4] S. Goferman, L. Zelnik-Manor, and A. Tal. Context-aware saliency detection. In IEEE CVPR, 2010. [5] Y. Guo et al. Image retargeting using mesh parametrization. IEEE Trans. Multimedia, 2009. [6] L. Liu, R. Chen, L. Wolf, and D. Cohen-Or. Optimizing photo composition. Computer Graphics Forum, 2010. [7] M. Nishiyama, T. Okabe, Y. Sato, and I. Sato. Sensation-based photo cropping. In ACM Multimedia, 2009. [8] A. Santella et al. Gaze-based interaction for semi-automatic photo cropping. In SIGCHI, 2006. [9] L.-K. Wong and K.-L. Wong. Enhancing visual dominance by semantics-preserving image recomposition. In ACM Multimedia, 2012. [10] F.-L. Zhang, M. Wang, and S.-M. Hu. Aesthetic image enhance- ment by dependence-aware object recomposition. IEEE Trans. Multimedia, 2013. 930