Skip navigation
Run Run Shaw Library City University of Hong KongRun Run Shaw Library

Please use this identifier to cite or link to this item: http://dspace.cityu.edu.hk/handle/2031/7897
Full metadata record
DC FieldValueLanguage
dc.contributor.authorYao, Ting (姚霆)en_US
dc.date.accessioned2015-09-25T02:08:08Z
dc.date.accessioned2017-09-19T09:19:39Z
dc.date.accessioned2019-02-12T08:41:21Z-
dc.date.available2015-09-25T02:08:08Z
dc.date.available2017-09-19T09:19:39Z
dc.date.available2019-02-12T08:41:21Z-
dc.date.issued2014en_US
dc.identifier.othercs2015-004en_US
dc.identifier.urihttp://144.214.8.231/handle/2031/7897-
dc.descriptionCityU Call Number: QA76.575 .Y37 2014en_US
dc.descriptionxiii, 147 leaves : ill. 30 cm.en_US
dc.descriptionThesis (Ph.D.)--City University of Hong Kong, 2014.en_US
dc.descriptionIncludes bibliographical references (leaves 134-145)en_US
dc.description.abstractThis thesis investigates the problem of multimedia search under the umbrella of knowledge transfer by considering three cases: 1) how to exploit visual patterns from the initial ranked list to boost search precision, 2) how to leverage the external knowledge as a prior to help the search, and 3) how to explore the largely available click-through data (i.e., crowdsourcing human intelligence) for annotation and search. A common practice for improving search performance is to rerank the initial visual documents returned from a search engine by seeking consensus from various visual features. We propose a new reranking algorithm, named circular reranking, that reinforces the mutual exchange of information across multiple modalities for improving search performance, following the philosophy that strong performing modality could learn from weaker ones, while weak modality does benefit from interacting with stronger ones. Technically, circular reranking conducts multiple runs of random walks through exchanging the ranking scores among different features in a cyclic manner. Moreover, we study several properties of circular reranking, including how and which order of information propagation should be configured to fully exploit the potential of modalities for reranking. For the transfer of external knowledge, we first systematically analyze the different factors that lead to the success and failure of transferring classifiers. A simple yet innovative and practical model is proposed for predicting the transfer from the clues such as the distribution shift of data, concept category and concept contextual relationship. Next, we develop the semi-supervised domain adaptation with subspace learning and transfer RankBoost algorithms for one-to-one domain adaptation and multiple-to-one domain adaptation, respectively. The former aims to jointly explore invariant low-dimensional structures across domains to correct data distribution mismatch and leverage available unlabeled target examples to exploit the underlying intrinsic information in the target domain. The later extends the generic RankBoost learning framework for transferring knowledge from multiple sources. To investigate the use of click-through data, we devise a novel video similarity measurement based on polynomial semantic indexing. Two mappings to project queries and video documents into a common latent space are learnt by minimizing the margin ranking loss of the observed query-video pairs on the click-through bipartite. Then the dot product in the latent space is taken as the similarity function between videos and the video similarity is further applied for three major tasks in video tagging: tag assignment, ranking, and enrichment. Later, to bridge the user intention gap and allow direct comparison of text queries and visual images, click-through-based cross-view learning approach is presented for image search. The objective is formalized as a latent space learning by jointly minimizing the distance between the mappings of query and image in the latent space and preserving the inherent structure in each original space. We evaluate all the proposed techniques on several large-scale real-world image and video datasets. Experimental evaluations demonstrate promising results of our techniques, and their advantages to be applied to various multimedia search applications.en_US
dc.publisherCity University of Hong Kongen_US
dc.rightsThis work is protected by copyright. Reproduction or distribution of the work in any format is prohibited without written permission of the copyright owner.en_US
dc.rightsAccess is unrestricted.en_US
dc.subject.lcshMultimedia systems.en_US
dc.subject.lcshDatabase searching.en_US
dc.titleMultimedia search by self, external, and crowdsourcing knowledgeen_US
dc.title.alternativeDuo mei ti sou suo : cong yuan shi dui xiang, xiang guan zi yuan dao qun ti zhi hui fen xien_US
dc.title.alternative多媒體搜索 : 從原始對象, 相關資源到群體智慧分析en_US
dc.typethesisen_US
dc.contributor.departmentDepartment of Computer Scienceen_US
dc.degree.disciplineComputer Scienceen_US
dc.degree.leveldoctoralen_US
dc.degree.nameDoctor of Philosophyen_US
dc.description.awardWon the 2015 SIGMM Outstanding Ph.D. Thesis Award.en_US
dc.description.fulltextAward winning work is available.en_US
dc.identifier.cataloghttp://lib.cityu.edu.hk/record=b4693549en_US
Appears in Collections:Student Works With External Awards 

Files in This Item:
File Description SizeFormat 
abstract.html132 BHTMLView/Open
fulltext.html132 BHTMLView/Open
award_news.html123 BHTMLView/Open
Show simple item record


Items in Digital CityU Collections are protected by copyright, with all rights reserved, unless otherwise indicated.

Send feedback to Library Systems
Privacy Policy | Copyright | Disclaimer