TY - GEN
T1 - High-level semantic photographic composition analysis and understanding with deep neural networks
AU - Wu, Min Tzu
AU - Pan, Tse Yu
AU - Tsai, Wan Lun
AU - Kuo, Hsu Chan
AU - Hu, Min Chun
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/9/5
Y1 - 2017/9/5
N2 - In order to take better photos, it is a fundamental step for the beginners of photography to learn basic photo composition rules. However, there are no tools developed to help beginners analyze the composition rules in given photos. Thus, in this study we developed a system with the capability to identify 12 common composition rules in a photo. It should be noted that some of the 12 common composition rules have not been considered by the previous studies, and this deficit gives this study its significance and appropriateness. In particular, we utilized deep neural networks (DNN) to extract high-level semantic features for facilitating the further analysis of photo composition rules. In order to train the DNN model, our research team constructed a dataset, which is collected from some famous photo websites, such as DPChallenge, Flicker, and Unsplash. All the collected photos were later labelled with 12 composition rules by a wide range of raters recruited from Amazon Mechanical Turk (AMT). Two DNN architectures (AlexNet and GoogLeNet) were then employed to build our system based on the collected dataset. The representative features of each composition rule were further visualized in our system. The results showed the feasibility of the proposed system and revealed the possibility of using this system to assist potential users to improve their photographical skills and expertise.
AB - In order to take better photos, it is a fundamental step for the beginners of photography to learn basic photo composition rules. However, there are no tools developed to help beginners analyze the composition rules in given photos. Thus, in this study we developed a system with the capability to identify 12 common composition rules in a photo. It should be noted that some of the 12 common composition rules have not been considered by the previous studies, and this deficit gives this study its significance and appropriateness. In particular, we utilized deep neural networks (DNN) to extract high-level semantic features for facilitating the further analysis of photo composition rules. In order to train the DNN model, our research team constructed a dataset, which is collected from some famous photo websites, such as DPChallenge, Flicker, and Unsplash. All the collected photos were later labelled with 12 composition rules by a wide range of raters recruited from Amazon Mechanical Turk (AMT). Two DNN architectures (AlexNet and GoogLeNet) were then employed to build our system based on the collected dataset. The representative features of each composition rule were further visualized in our system. The results showed the feasibility of the proposed system and revealed the possibility of using this system to assist potential users to improve their photographical skills and expertise.
UR - http://www.scopus.com/inward/record.url?scp=85031705095&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85031705095&partnerID=8YFLogxK
U2 - 10.1109/ICMEW.2017.8026247
DO - 10.1109/ICMEW.2017.8026247
M3 - Conference contribution
AN - SCOPUS:85031705095
T3 - 2017 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2017
SP - 279
EP - 284
BT - 2017 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2017
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2017 IEEE International Conference on Multimedia and Expo Workshops, ICMEW 2017
Y2 - 10 July 2017 through 14 July 2017
ER -