Open Access. Powered by Scholars. Published by Universities.®

Databases and Information Systems Commons

Open Access. Powered by Scholars. Published by Universities.®

OS and Networks

Research Collection School Of Computing and Information Systems

Multi-modal

Publication Year

Articles 1 - 2 of 2

Full-Text Articles in Databases and Information Systems

Learning Semantically Rich Network-Based Multi-Modal Mobile User Interface Embeddings, Meng Kiat Gary Ang, Ee-Peng Lim May 2022

Learning Semantically Rich Network-Based Multi-Modal Mobile User Interface Embeddings, Meng Kiat Gary Ang, Ee-Peng Lim

Research Collection School Of Computing and Information Systems

Semantically rich information from multiple modalities - text, code, images, categorical and numerical data - co-exist in the user interface (UI) design of mobile applications. Moreover, each UI design is composed of inter-linked UI entities which support different functions of an application, e.g., a UI screen comprising a UI taskbar, a menu and multiple button elements. Existing UI representation learning methods unfortunately are not designed to capture multi-modal and linkage structure between UI entities. To support effective search and recommendation applications over mobile UIs, we need UI representations that integrate latent semantics present in both multi-modal information and linkages between …


Learning Network-Based Multi-Modal Mobile User Interface Embeddings, Gary Ang, Ee-Peng Lim Apr 2021

Learning Network-Based Multi-Modal Mobile User Interface Embeddings, Gary Ang, Ee-Peng Lim

Research Collection School Of Computing and Information Systems

Rich multi-modal information - text, code, images, categorical and numerical data - co-exist in the user interface (UI) design of mobile applications. UI designs are composed of UI entities supporting different functions which together enable the application. To support effective search and recommendation applications over mobile UIs, we need to be able to learn UI representations that integrate latent semantics. In this paper, we propose a novel unsupervised model - Multi-modal Attention-based Attributed Network Embedding (MAAN) model. MAAN is designed to capture both multi-modal and structural network information. Based on the encoder-decoder framework, MAAN aims to learn UI representations that …