Open Access. Powered by Scholars. Published by Universities.®

Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

Articles 1 - 8 of 8

Full-Text Articles in Engineering

Topical Summarization Of Web Videos By Visual-Text Time-Dependent Alignment, Song Tan, Hung-Khoon Tan, Chong-Wah Ngo Dec 2010

Topical Summarization Of Web Videos By Visual-Text Time-Dependent Alignment, Song Tan, Hung-Khoon Tan, Chong-Wah Ngo

Research Collection School Of Computing and Information Systems

Search engines are used to return a long list of hundreds or even thousands of videos in response to a query topic. Efficient navigation of videos becomes difficult and users often need to painstakingly explore the search list for a gist of the search result. This paper addresses the challenge of topical summarization by providing a timeline-based visualization of videos through matching of heterogeneous sources. To overcome the so called sparse-text problem of web videos, auxiliary information from Google context is exploited. Google Trends is used to predict the milestone events of a topic. Meanwhile, the typical scenes of web …


Data-Driven Approaches To Community-Contributed Video Applications, Xiao Wu, Chong-Wah Ngo, Wan-Lei Zhao Oct 2010

Data-Driven Approaches To Community-Contributed Video Applications, Xiao Wu, Chong-Wah Ngo, Wan-Lei Zhao

Research Collection School Of Computing and Information Systems

With the prosperity of video-sharing websites such as YouTube, the amount of community-contributed video has increased dramatically. Reportedly more than 65,000 new videos were uploaded to YouTube every day in July 2006 and it's estimated that 20 hours of new videos were uploaded to the site every minute in May 2009. In addition to the huge volume of video data, the social Web provides rich contextual and social resources associated with videos. These resources include title, tags, thumbnails, related videos, and user and community information, as illustrated in Figure 1. While billions of user-generated videos accompanied with rich-media information have …


On The Annotation Of Web Videos By Efficient Near-Duplicate Search, Wan-Lei Zhao, Xiao Wu, Chong-Wah Ngo Aug 2010

On The Annotation Of Web Videos By Efficient Near-Duplicate Search, Wan-Lei Zhao, Xiao Wu, Chong-Wah Ngo

Research Collection School Of Computing and Information Systems

With the proliferation of Web 2.0 applications, usersupplied social tags are commonly available in social media as a means to bridge the semantic gap. On the other hand, the explosive expansion of social web makes an overwhelming number of web videos available, among which there exists a large number of near-duplicate videos. In this paper, we investigate techniques which allow effective annotation of web videos from a data-driven perspective. A novel classifier-free video annotation framework is proposed by first retrieving visual duplicates and then suggesting representative tags. The significance of this paper lies in the addressing of two timely issues …


Investigating Perceptions Of A Location-Based Annotation System, Huynh Nhu Hop Quach, Khasfariyati Razikin, Dion Hoe-Lian Goh, Thi Nhu Quynh Kim, Tan Phat Pham, Yin-Leng Theng, Ee-Peng Lim Aug 2010

Investigating Perceptions Of A Location-Based Annotation System, Huynh Nhu Hop Quach, Khasfariyati Razikin, Dion Hoe-Lian Goh, Thi Nhu Quynh Kim, Tan Phat Pham, Yin-Leng Theng, Ee-Peng Lim

Research Collection School Of Computing and Information Systems

We introduce MobiTOP, a Web-based system for organizing and retrieving hierarchical location-based annotations. Each annotation contains multimedia content (such as text, images, video) associated with a location, and users are able to annotate existing annotations to an arbitrary depth, in effect creating a hierarchy. An evaluation was conducted on a group of potential users to ascertain their perceptions of the usability of the application. The results were generally positive and the majority of the participants saw MobiTOP as a useful platform to share location-based information. We conclude with implications of our work and opportunities for future research.


Coherent Bag-Of Audio Words Model For Efficient Large-Scale Video Copy Detection, Yang Liu, Wan-Lei Zhao, Chong-Wah Ngo, Chang-Sheng Xu, Han-Qing Lu Jul 2010

Coherent Bag-Of Audio Words Model For Efficient Large-Scale Video Copy Detection, Yang Liu, Wan-Lei Zhao, Chong-Wah Ngo, Chang-Sheng Xu, Han-Qing Lu

Research Collection School Of Computing and Information Systems

Current content-based video copy detection approaches mostly concentrate on the visual cues and neglect the audio information. In this paper, we attempt to tackle the video copy detection task resorting to audio information, which is equivalently important as well as visual information in multimedia processing. Firstly, inspired by bag-of visual words model, a bag-of audio words (BoA) representation is proposed to characterize each audio frame. Different from naive singlebased modeling audio retrieval approaches, BoA is a highlevel model due to its perceptual and semantical property. Within the BoA model, a coherency vocabulary indexing structure is adopted to achieve more efficient …


Co-Reranking By Mutual Reinforcement For Image Search, Ting Yao, Tao Mei, Chong-Wah Ngo Jul 2010

Co-Reranking By Mutual Reinforcement For Image Search, Ting Yao, Tao Mei, Chong-Wah Ngo

Research Collection School Of Computing and Information Systems

Most existing reranking approaches to image search focus solely on mining “visual” cues within the initial search results. However, the visual information cannot always provide enough guidance to the reranking process. For example, different images with similar appearance may not always present the same relevant information to the query. Observing that multi-modality cues carry complementary relevant information, we propose the idea of co-reranking for image search, by jointly exploring the visual and textual information. Co-reranking couples two random walks, while reinforcing the mutual exchange and propagation of information relevancy across different modalities. The mutual reinforcement is iteratively updated to constrain …


On The Sampling Of Web Images For Learning Visual Concept Classifiers, Shiai Zhu, Gang Wang, Chong-Wah Ngo, Yu-Gang Jiang Jul 2010

On The Sampling Of Web Images For Learning Visual Concept Classifiers, Shiai Zhu, Gang Wang, Chong-Wah Ngo, Yu-Gang Jiang

Research Collection School Of Computing and Information Systems

Visual concept learning often requires a large set of training images. In practice, nevertheless, acquiring noise-free training labels with sufficient positive examples is always expensive. A plausible solution for training data collection is by sampling the largely available user-tagged images from social media websites. With the general belief that the probability of correct tagging is higher than that of incorrect tagging, such a solution often sounds feasible, though is not without challenges. First, user-tags can be subjective and, to certain extent, are ambiguous. For instance, an image tagged with “whales” may be simply a picture about ocean museum. Learning concept …


Pagesense: Style-Wise Web Page Advertising, Lusong Li, Tao Mei, Xiang Niu, Chong-Wah Ngo Apr 2010

Pagesense: Style-Wise Web Page Advertising, Lusong Li, Tao Mei, Xiang Niu, Chong-Wah Ngo

Research Collection School Of Computing and Information Systems

This paper presents an innovative style-wise advertising platform for web page. Web page “style” mainly refers to visual effects, such as color and layout. Unlike the most popular ad-network such as Google AdSense which needs publishers to change the original structure of their pages and define the position and style of the embedded ads manually, stylewise page advertising aims to automatically deliver styleconsistent ads at proper positions within the web page, without breaking the layout of the original page. Our system is motivated from the fact that almost 90% web pages contain blank regions without any content. Given a web …