<XML><RECORDS><RECORD><REFERENCE_TYPE>3</REFERENCE_TYPE><REFNUM>9195</REFNUM><AUTHORS><AUTHOR>Ren,R.</AUTHOR><AUTHOR>Halvey,M.</AUTHOR><AUTHOR>Jose,J.M.</AUTHOR></AUTHORS><YEAR>2009</YEAR><TITLE>Audio-visual Feature Aggregation for Query Generation</TITLE><PLACE_PUBLISHED>ICME 2009</PLACE_PUBLISHED><PUBLISHER>IEEE Computer Society Press</PUBLISHER><ISBN>978-988-17012-5-1</ISBN><LABEL>Ren:2009:9195</LABEL><KEYWORDS><KEYWORD>feature aggregation</KEYWORD></KEYWORDS<ABSTRACT>Using multiple examples has become a popular query scenario in multimedia retrieval. This paper explores a unified representation which accumulates various features from different examples to denote a query. Continuous low-level features are quantised into a set of discrete variants. These variants follow a similar distribution as text terms do in a given document collection. Three criteria are compared to justify this projection, including minimised chi-square, maximised entropy and minimised AC/DC. Statistics similar to text term frequency are computed from these variants for document similarity ranking. Two ranking functions, KL divergence and BM25, are used for multimedia retrieval. The evaluation collection consists of the Corel image set and TRECVid 2006 collection with four low-level visual features. Experimental results show that the overall query performance based on this representation is comparable and in some cases out-performs direct visual feature comparison and the K-median clustering.</ABSTRACT></RECORD></RECORDS></XML>