Saha B, Pham D-S, Phung D and Venkatesh S (2013), "Clustering patient medical records via sparse subspace representation", In Pacific-Asia Conference on Knowledge Discovery and Data Mining. , pp. 123-134. Springer. |
Abstract: The health industry is facing increasing challenge with “big data” as traditional methods fail to manage the scale and complexity. This paper examines clustering of patient records for chronic diseases to facilitate a better construction of care plans. We solve this problem under the framework of subspace clustering. Our novel contribution lies in the exploitation of sparse representation to discover subspaces automatically and a domain-specific construction of weighting matrices for patient records. We show the new formulation is readily solved by extending existing L1-regularized optimization algorithms. Using a cohort of both diabetes and stroke data we show that we outperform existing benchmark clustering techniques in the literature. |
BibTeX:
@inproceedings{budhaditya2013clustering, author = {Saha, Budhaditya and Pham, Duc-Son and Phung, Dinh and Venkatesh, Svetha}, title = {Clustering patient medical records via sparse subspace representation}, booktitle = {Pacific-Asia Conference on Knowledge Discovery and Data Mining}, publisher = {Springer}, year = {2013}, pages = {123--134}, url = {http://link.springer.com/chapter/10.1007/978-3-642-37456-2_11} } |
Saha B, Pham DS, Phung D and Venkatesh S (2013), "Sparse Subspace Clustering via Group Sparse Coding", In Thirteenth SIAM International Conference on Data Mining. , pp. 130-138. |
Abstract: We propose in this paper a novel sparse subspace clustering method that regularizes sparse subspace representation by exploiting the structural sharing between tasks and data points via group sparse coding. We derive simple, provably convergent, and computationally efficient algorithms for solving the proposed group formulations. We demonstrate the advantage of the framework on three challenging benchmark datasets ranging from medical record data to image and text clustering and show that they consistently outperforms rival methods. |
BibTeX:
@inproceedings{budhaditya2013sparse, author = {Saha, Budhaditya and Pham, Duc Son and Phung, Dinh and Venkatesh, Svetha}, title = {Sparse Subspace Clustering via Group Sparse Coding}, booktitle = {Thirteenth SIAM International Conference on Data Mining}, year = {2013}, pages = {130--138}, url = {http://epubs.siam.org/doi/abs/10.1137/1.9781611972832.15} } |
Gupta S, Phung D and Venkatesh S (2013), "Factorial Multi-Task Learning: A Bayesian Nonparametric Approach", In 30th International Conference on Machine Learning (ICML). , pp. 657-665. |
Abstract: Multi-task learning is a paradigm shown to improve the performance of related tasks through their joint learning. However, for real-world data, it is usually difficult to assess the task relatedness and joint learning with unrelated tasks may lead to serious performance degradations. To this end, we propose a framework that groups the tasks based on their relatedness in a low dimensional subspace and allows a varying degree of relatedness among tasks by sharing the subspace bases across the groups. This provides the flexibility of no sharing when two sets of tasks are unrelated and partial/total sharing when the tasks are related. Importantly, the number of task-groups and the subspace dimensionality are automatically inferred from the data. This feature keeps the model beyond a specific set of parameters. To realize our framework, we present a novel Bayesian nonparametric prior that extends the traditional hierarchical beta process prior using a Dirichlet process to permit potentially infinite number of child beta processes. We apply our model for multi-task regression and classification applications. Experimental results using several synthetic and real-world datasets show the superiority of our model to other recent state-of-the-art multi-task learning methods. |
BibTeX:
@inproceedings{gupta2013factorial, author = {Gupta, Sunil and Phung, Dinh and Venkatesh, Svetha}, title = {Factorial Multi-Task Learning: A Bayesian Nonparametric Approach}, booktitle = {30th International Conference on Machine Learning (ICML)}, year = {2013}, pages = {657--665}, url = {http://jmlr.org/proceedings/papers/v28/gupta13a.pdf} } |
Gupta S, Phung D, Adams B and Venkatesh S (2013), "Regularized nonnegative shared subspace learning", Data mining and knowledge discovery. Vol. 26(1), pp. 57-97. Springer. |
Abstract: Joint modeling of related data sources has the potential to improve various data mining tasks such as transfer learning, multitask clustering, information retrieval etc. However, diversity among various data sources might outweigh the advantages of the joint modeling, and thus may result in performance degradations. To this end, we propose a regularized shared subspace learning framework, which can exploit the mutual strengths of related data sources while being immune to the effects of the variabilities of each source. This is achieved by further imposing a mutual orthogonality constraint on the constituent subspaces which segregates the common patterns from the source specific patterns, and thus, avoids performance degradations. Our approach is rooted in nonnegative matrix factorization and extends it further to enable joint analysis of related data sources. Experiments performed using three real world data sets for both retrieval and clustering applications demonstrate the benefits of regularization and validate the effectiveness of the model. Our proposed solution provides a formal framework appropriate for jointly analyzing related data sources and therefore, it is applicable to a wider context in data mining. |
BibTeX:
@article{gupta2013regularized, author = {Gupta, Sunil and Phung, Dinh and Adams, Brett and Venkatesh, Svetha}, title = {Regularized nonnegative shared subspace learning}, journal = {Data mining and knowledge discovery}, publisher = {Springer}, year = {2013}, volume = {26}, number = {1}, pages = {57--97}, url = {http://link.springer.com/article/10.1007%2Fs10618-011-0244-8} } |
Li C, Phung D, Rana S and Venkatesh S (2013), "Exploiting side information in distance dependent Chinese restaurant processes for data clustering", In IEEE International Conference on Multimedia and Expo (ICME). , pp. 1-6. |
Abstract: Multimedia contents often possess weakly annotated data such as tags, links and interactions. The weakly annotated data is called side information. It is the auxiliary information of data and provides hints for exploring the link structure of data. Most clustering algorithms utilize pure data for clustering. A model that combines pure data and side information, such as images and tags, documents and keywords, can perform better at understanding the underlying structure of data. We demonstrate how to incorporate different types of side information into a recently proposed Bayesian nonparametric model, the distance dependent Chinese restaurant process (DD-CRP). Our algorithm embeds the affinity of this information into the decay function of the DD-CRP when side information is in the form of subsets of discrete labels. It is flexible to measure distance based on arbitrary side information instead of only the spatial layout or time stamp of observations. At the same time, for noisy and incomplete side information, we set the decay function so that the DD-CRP reduces to the traditional Chinese restaurant process, thus not inducing side effects of noisy and incomplete side information. Experimental evaluations on two real-world datasets NUS WIDE and 20 Newsgroups show exploiting side information in DD-CRP significantly improves the clustering performance. |
BibTeX:
@inproceedings{li2013exploiting, author = {Li, Cheng and Phung, Dinh and Rana, Santu and Venkatesh, Svetha}, title = {Exploiting side information in distance dependent Chinese restaurant processes for data clustering}, booktitle = {IEEE International Conference on Multimedia and Expo (ICME)}, year = {2013}, pages = {1--6}, url = {http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6607475} } |
Low SY, Pham DS and Venkatesh S (2013), "Compressive speech enhancement", Speech Communication. Vol. 55(6), pp. 757-768. Elsevier. |
Abstract: This paper presents an alternative approach to speech enhancement by using compressed sensing (CS). CS is a new sampling theory, which states that sparse signals can be reconstructed from far fewer measurements than the Nyquist sampling. As such, CS can be exploited to reconstruct only the sparse components (e.g., speech) from the mixture of sparse and non-sparse components (e.g., noise). This is possible because in a time-frequency representation, speech signal is sparse whilst most noise is non-sparse. Derivation shows that on average the signal to noise ratio (SNR) in the compressed domain is greater or equal than the uncompressed domain. Experimental results concur with the derivation and the proposed CS scheme achieves better or similar perceptual evaluation of speech quality (PESQ) scores and segmental SNR compared to other conventional methods in a wide range of input SNR. |
BibTeX:
@article{low2013compressive, author = {Low, Siow Yong and Pham, Duc Son and Venkatesh, Svetha}, title = {Compressive speech enhancement}, journal = {Speech Communication}, publisher = {Elsevier}, year = {2013}, volume = {55}, number = {6}, pages = {757--768}, url = {http://www.sciencedirect.com/science/article/pii/S0167639313000356} } |
Naseem I, Pham D-S and Venkatesh S (2013), "A novel information theoretic approach to wavelet feature selection for texture classification", Computers & Electrical Engineering. Vol. 39(2), pp. 319-325. Elsevier. |
Abstract: In this research we address the problem of discriminant subband selection for texture classification. A novel Effective Information based Subband Selection (EISS) algorithm is proposed which utilizes the intra-class and inter-class distributions. Essentially these distributions are used to calculate the class-based entropy for a given subband. This class-based information is incorporated in the total information content of the training images to develop a robust Effective Information (EI) criterion. Only the subbands with the top EI criteria are allowed to participate in the classification process. The proposed EISS algorithm is evaluated on Brodatz texture database and has shown to outperform the most relevant method based on mutual information criterion. |
BibTeX:
@article{naseem2013novel, author = {Naseem, Imran and Pham, Duc-Son and Venkatesh, Svetha}, title = {A novel information theoretic approach to wavelet feature selection for texture classification}, journal = {Computers & Electrical Engineering}, publisher = {Elsevier}, year = {2013}, volume = {39}, number = {2}, pages = {319--325}, url = {http://www.sciencedirect.com/science/article/pii/S0045790612002145} } |
Nguyen T, Phung D and Venkatesh S (2013), "Analysis of psycholinguistic processes and topics in online autism communities", In IEEE International Conference on Multimedia and Expo. , pp. 1-6. |
Abstract: Current growth of individuals on the autism spectrum disorder (ASD) requires continuous support and care. With the popularity of social media, online communities of people affected by ASD emerge. This paper presents an analysis of these online communities through understanding aspects that differentiate such communities. In this paper, the aspects given are not expressed in terms of friendship, exchange of information, social support or recreation, but rather with regard to the topics and linguistic styles that people express in their on-line writing. Using data collected unobtrusively from LiveJournal, we analyze posts made by ten autism communities in conjunction with those made by a control group of standard communities. Significant differences have been found between autism and control communities when characterized by latent topics of discussion and psycholinguistic features. Latent topics are found to have greater predictive power than linguistic features when classifying blog posts as either autism or control community. This study suggests that data mining of online blogs has the potential to detect clinically meaningful data. It opens the door to possibilities including sentinel risk surveillance and harnessing the power in diverse large datasets. |
BibTeX:
@inproceedings{Nguyen2013a, author = {Nguyen, Thin and Phung, Dinh and Venkatesh, Svetha}, title = {Analysis of psycholinguistic processes and topics in online autism communities}, booktitle = {IEEE International Conference on Multimedia and Expo}, year = {2013}, pages = {1--6}, url = {http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6607615} } |
Nguyen TD, Tran T, Phung D and Venkatesh S (2013), "Learning Parts-based Representations with Nonnegative Restricted Boltzmann Machine", In Asian Conference on Machine Learning. , pp. 133-148. |
Abstract: The success of any machine learning system depends critically on effective representations of data. In many cases, especially those in vision, it is desirable that a representation scheme uncovers the parts-based, additive nature of the data. Of current representation learning schemes, restricted Boltzmann machines (RBMs) have proved to be highly effective in unsupervised settings. However, when it comes to parts-based discovery, RBMs do not usually produce satisfactory results. We enhance such capacity of RBMs by introducing nonnegativity into the model weights, resulting in a variant called nonnegative restricted Boltzmann machine (NRBM). The NRBM produces not only controllable decomposition of data into interpretable parts but also offers a way to estimate the intrinsic nonlinear dimensionality of data. We demonstrate the capacity of our model on well-known datasets of handwritten digits, faces and documents. The decomposition quality on images is comparable with or better than what produced by the nonnegative matrix factorisation (NMF), and the thematic features uncovered from text are qualitatively interpretable in a similar manner to that of the latent Dirichlet allocation (LDA). However, the learnt features, when used for classification, are more discriminative than those discovered by both NMF and LDA and comparable with those by RBM. |
BibTeX:
@inproceedings{Nguyen2013b, author = {Nguyen, Tu Dinh and Tran, Truyen and Phung, Dinh and Venkatesh, Svetha}, title = {Learning Parts-based Representations with Nonnegative Restricted Boltzmann Machine}, booktitle = {Asian Conference on Machine Learning}, year = {2013}, pages = {133--148}, url = {http://jmlr.org/proceedings/papers/v29/Nguyen13.pdf} } |
Nguyen T, Phung D, Adams B and Venkatesh S (2013), "Event extraction using behaviors of sentiment signals and burst structure in social media", Knowledge and information systems. Vol. 37(2), pp. 279-304. Springer. |
Abstract: Significant world events often cause the behavioral convergence of the expression of shared sentiment. This paper examines the use of the blogosphere as a framework to study user psychological behaviors, using their sentiment responses as a form of ‘sensor’ to infer real-world events of importance automatically. We formulate a novel temporal sentiment index function using quantitative measure of the valence value of bearing words in blog posts in which the set of affective bearing words is inspired from psychological research in emotion structure. The annual local minimum and maximum of the proposed sentiment signal function are utilized to extract significant events of the year and corresponding blog posts are further analyzed using topic modeling tools to understand their content. The paper then examines the correlation of topics discovered in relation to world news events reported by the mainstream news service provider, Cable News Network, and by using the Google search engine. Next, aiming at understanding sentiment at a finer granularity over time, we propose a stochastic burst detection model, extended from the work of Kleinberg, to work incrementally with stream data. The proposed model is then used to extract sentimental bursts occurring within a specific mood label (for example, a burst of observing ‘shocked’). The blog posts at those time indices are analyzed to extract topics, and these are compared to real-world news events. Our comprehensive set of experiments conducted on a large-scale set of 12 million posts from Livejournal shows that the proposed sentiment index function coincides well with significant world events while bursts in sentiment allow us to locate finer-grain external world events. |
BibTeX:
@article{nguyen2013event, author = {Nguyen, Thin and Phung, Dinh and Adams, Brett and Venkatesh, Svetha}, title = {Event extraction using behaviors of sentiment signals and burst structure in social media}, journal = {Knowledge and information systems}, publisher = {Springer}, year = {2013}, volume = {37}, number = {2}, pages = {279--304}, url = {http://prada-research.net/ dinh/uploads/Main/Publications/Nguyen_etal_kais12.pdf} } |
Nguyen T, Phung D, Gupta S and Venkatesh S (2013), "Extraction of latent patterns and contexts from social honest signals using hierarchical Dirichlet processes", In IEEE International Conference on Pervasive Computing and Communications. , pp. 47-55. |
Abstract: A fundamental task in pervasive computing is reliable acquisition of contexts from sensor data. This is crucial to the operation of smart pervasive systems and services so that they might behave efficiently and appropriately upon a given context. Simple forms of context can often be extracted directly from raw data. Equally important, or more, is the hidden context and pattern buried inside the data, which is more challenging to discover. Most of existing approaches borrow methods and techniques from machine learning, dominantly employ parametric unsupervised learning and clustering techniques. Being parametric, a severe drawback of these methods is the requirement to specify the number of latent patterns in advance. In this paper, we explore the use of Bayesian nonparametric methods, a recent data modelling framework in machine learning, to infer latent patterns from sensor data acquired in a pervasive setting. Under this formalism, nonparametric prior distributions are used for data generative process, and thus, they allow the number of latent patterns to be learned automatically and grow with the data - as more data comes in, the model complexity can grow to explain new and unseen patterns. In particular, we make use of the hierarchical Dirichlet processes (HDP) to infer atomic activities and interaction patterns from honest signals collected from sociometric badges. We show how data from these sensors can be represented and learned with HDP. We illustrate insights into atomic patterns learned by the model and use them to achieve high-performance clustering. We also demonstrate the framework on the popular Reality Mining dataset, illustrating the ability of the model to automatically infer typical social groups in this dataset. Finally, our framework is generic and applicable to a much wider range of problems in pervasive computing where one needs to infer high-level, latent patterns and contexts from sensor data. |
BibTeX:
@inproceedings{nguyen2013extraction, author = {Nguyen, Thuong and Phung, Dinh and Gupta, Sunil and Venkatesh, Svetha}, title = {Extraction of latent patterns and contexts from social honest signals using hierarchical Dirichlet processes}, booktitle = {IEEE International Conference on Pervasive Computing and Communications}, year = {2013}, pages = {47--55}, url = {http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6526713} } |
Nguyen T-V, Phung D, Gupta S and Venkatesh S (2013), "Interactive browsing system for anomaly video surveillance", In IEEE Eighth International Conference on Intelligent Sensors, Sensor Networks and Information Processing. , pp. 384-389. |
Abstract: Existing anomaly detection methods in video surveillance exhibit lack of congruence between rare events detected by algorithms and what is considered anomalous by users. This paper introduces a novel browsing model to address this issue, allowing users to interactively examine rare events in an intuitive manner. Introducing a novel way to compute rare motion patterns, we estimate latent factors of foreground motion patterns through Bayesian Nonparametric Factor analysis. Each factor corresponds to a typical motion pattern. A rarity score for each factor is computed, and ordered in decreasing order of rarity, permitting users to browse events using any proportion of rare factors. Rare events correspond to frames that contain the rare factors chosen. We present the user with an interface to inspect events that incorporate these rarest factors in a spatial-temporal manner. We demonstrate the system on a public video data set, showing key aspects of the browsing paradigm. |
BibTeX:
@inproceedings{nguyen2013interactive, author = {Nguyen, Tien-Vu and Phung, Dinh and Gupta, Sunil and Venkatesh, Svetha}, title = {Interactive browsing system for anomaly video surveillance}, booktitle = {IEEE Eighth International Conference on Intelligent Sensors, Sensor Networks and Information Processing}, year = {2013}, pages = {384--389}, url = {http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6529821} } |
Nguyen TD, Tran T, Phung D and Venkatesh S (2013), "Latent Patient Profile Modelling and Applications with Mixed-Variate Restricted Boltzmann Machine", In Pacific-Asia Conference on Knowledge Discovery and Data Mining. , pp. 123-135. Springer. |
Abstract: Efficient management of chronic diseases is critical in modern health care. We consider diabetes mellitus, and our ongoing goal is to examine how machine learning can deliver information for clinical efficiency. The challenge is to aggregate highly heterogeneous sources including demographics, diagnoses, pathologies and treatments, and extract similar groups so that care plans can be designed. To this end, we extend our recent model, the mixed-variate restricted Boltzmann machine (MV.RBM), as it seamlessly integrates multiple data types for each patient aggregated over time and outputs a homogeneous representation called “latent profile” that can be used for patient clustering, visualisation, disease correlation analysis and prediction. We demonstrate that the method outperforms all baselines on these tasks - the primary characteristics of patients in the same groups are able to be identified and the good result can be achieved for the diagnosis codes prediction. |
BibTeX:
@inproceedings{nguyen2013latent, author = {Nguyen, Tu Dinh and Tran, Truyen and Phung, Dinh and Venkatesh, Svetha}, title = {Latent Patient Profile Modelling and Applications with Mixed-Variate Restricted Boltzmann Machine}, booktitle = {Pacific-Asia Conference on Knowledge Discovery and Data Mining}, publisher = {Springer}, year = {2013}, pages = {123--135}, url = {http://prada-research.net/ svetha/papers/2013/tu_truyen_phung_venkatesh_pakdd13.pdf} } |
Nguyen TD, Tran T, Phung D and Venkatesh S (2013), "Learning sparse latent representation and distance metric for image retrieval", In IEEE International Conference on Multimedia and Expo. , pp. 1-6. |
Abstract: The performance of image retrieval depends critically on the semantic representation and the distance function used to estimate the similarity of two images. A good representation should integrate multiple visual and textual (e.g., tag) features and offer a step closer to the true semantics of interest (e.g., concepts). As the distance function operates on the representation, they are interdependent, and thus should be addressed at the same time. We propose a probabilistic solution to learn both the representation from multiple feature types and modalities and the distance metric from data. The learning is regularised so that the learned representation and information-theoretic metric will (i) preserve the regularities of the visual/textual spaces, (ii) enhance structured sparsity, (iii) encourage small intra-concept distances, and (iv) keep inter-concept images separated. We demonstrate the capacity of our method on the NUS-WIDE data. For the well-studied 13 animal subset, our method outperforms state-of-the-art rivals. On the subset of single-concept images, we gain 79:5% improvement over the standard nearest neighbours approach on the MAP score, and 45.7% on the NDCG. |
BibTeX:
@inproceedings{nguyen2013learning, author = {Nguyen, Tu Dinh and Tran, Truyen and Phung, Dinh and Venkatesh, Svetha}, title = {Learning sparse latent representation and distance metric for image retrieval}, booktitle = {IEEE International Conference on Multimedia and Expo}, year = {2013}, pages = {1--6}, url = {http://prada-research.net/ svetha/papers/2013/icme13_142.pdf} } |
Nguyen T, Dao B, Phung D, Venkatesh S, Berk M and others (2013), "Online Social Capital: Mood, Topical and Psycholinguistic Analysis.", In ICWSM: International Conference on Weblogs and Social Media. , pp. 449-456. |
Abstract: Social media provides rich sources of personal information and community interaction which can be linked to aspect of mental health. In this paper we investigate manifest properties of textual messages, including latent topics, psycholinguistic features, and authors' mood, of a large corpus of blog posts, to analyze the aspect of social capital in social media communities. Using data collected from Live Journal, we find that bloggers with lower social capital have fewer positive moods and more negative moods than those with higher social capital. It is also found that people with low social capital have more random mood swings over time than the people with high social capital. Significant differences are found between low and high social capital groups when characterized by a set of latent topics and psycholinguistic features derived from blogposts, suggesting discriminative features, proved to be useful for classification tasks. Good prediction is achieved when classifying among social capital groups using topic and linguistic features, with linguistic features are found to have greater predictive power than latent topics. The significance of our work lies in the importance of online social capital to potential construction of automatic healthcare monitoring systems. We further establish the link between mood and social capital in online communities, suggesting the foundation of new systems to monitor online mental well-being. |
BibTeX:
@inproceedings{nguyen2013online, author = {Nguyen, Thin and Dao, Bo and Phung, Dinh and Venkatesh, Svetha and Berk, Michael and others}, title = {Online Social Capital: Mood, Topical and Psycholinguistic Analysis.}, booktitle = {ICWSM: International Conference on Weblogs and Social Media}, year = {2013}, pages = {449--456}, url = {http://prada-research.net/ dinh/uploads/Main/Publications/nguyen_dao_phung_venkatesh_berk_icwsm13.pdf} } |
Nguyen T-V, Phung D and Venkatesh S (2013), "Topic model kernel: An empirical study towards probabilistically reduced features for classification", In Neural Information Processing. , pp. 124-131. |
Abstract: Probabilistic topic models have become a standard in modern machine learning with wide applications in organizing and summarizing ‘documents’ in high-dimensional data such as images, videos, texts, gene expression data, and so on. Representing data by dimensional reduction of mixture proportion extracted from topic models is not only richer in semantics than bag-of-word interpretation, but also more informative for classification tasks. This paper describes the Topic Model Kernel (TMK), a high dimensional mapping for Support Vector Machine classification of data generated from probabilistic topic models. The applicability of our proposed kernel is demonstrated in several classification tasks from real world datasets. We outperform existing kernels on the distributional features and give the comparative results on non-probabilistic data types. |
BibTeX:
@inproceedings{nguyen2013topic, author = {Nguyen, Tien-Vu and Phung, Dinh and Venkatesh, Svetha}, title = {Topic model kernel: An empirical study towards probabilistically reduced features for classification}, booktitle = {Neural Information Processing}, year = {2013}, pages = {124--131}, url = {http://prada-research.net/ dinh/uploads/Main/Publications/nguyen_etal_tr13.pdf} } |
Pham D-S, Saha B, Phung D and Venkatesh S (2013), "Detection of cross-channel anomalies", Knowledge and information systems. Vol. 35(1), pp. 33-59. Springer. |
Abstract: The data deluge has created a great challenge for data mining applications wherein the rare topics of interest are often buried in the flood of major headlines. We identify and formulate a novel problem: cross-channel anomaly detection from multiple data channels. Cross-channel anomalies are common among the individual channel anomalies and are often portent of significant events. Central to this new problem is a development of theoretical foundation and methodology. Using the spectral approach, we propose a two-stage detection method: anomaly detection at a single-channel level, followed by the detection of cross-channel anomalies from the amalgamation of single-channel anomalies. We also derive the extension of the proposed detection method to an online settings, which automatically adapts to changes in the data over time at low computational complexity using incremental algorithms. Our mathematical analysis shows that our method is likely to reduce the false alarm rate by establishing theoretical results on the reduction of an impurity index. We demonstrate our method in two applications: document understanding with multiple text corpora and detection of repeated anomalies in large-scale video surveillance. The experimental results consistently demonstrate the superior performance of our method compared with related state-of-art methods, including the one-class SVM and principal component pursuit. In addition, our framework can be deployed in a decentralized manner, lending itself for large-scale data stream analysis. |
BibTeX:
@article{pham2013detection, author = {Pham, Duc-Son and Saha, Budhaditya and Phung, Dinh and Venkatesh, Svetha}, title = {Detection of cross-channel anomalies}, journal = {Knowledge and information systems}, publisher = {Springer}, year = {2013}, volume = {35}, number = {1}, pages = {33--59}, url = {http://prada-research.net/ dinh/uploads/Main/Publications/Pham_etal_icdm11.pdf} } |
Pham D-S and Venkatesh S (2013), "Efficient algorithms for robust recovery of images from compressed data", IEEE Transactions on Image Processing. Vol. 22(12), pp. 4724-4737. IEEE. |
Abstract: Compressed sensing (CS) is an important theory for sub-Nyquist sampling and recovery of compressible data. Recently, it has been extended to cope with the case where corruption to the CS data is modeled as impulsive noise. The new formulation, termed as robust CS, combines robust statistics and CS into a single framework to suppress outliers in the CS recovery. To solve the newly formulated robust CS problem, a scheme that iteratively solves a number of CS problems-the solutions from which provably converge to the true robust CS solution-is suggested. This scheme is, however, rather inefficient as it has to use existing CS solvers as a proxy. To overcome limitations with the original robust CS algorithm, we propose in this paper more computationally efficient algorithms by following latest advances in large-scale convex optimization for nonsmooth regularization. Furthermore, we also extend the robust CS formulation to various settings, including additional affine constraints, l1-norm loss function, mix-norm regularization, and multitasking, so as to further improve robust CS and derive simple but effective algorithms to solve these extensions. We demonstrate that the new algorithms provide much better computational advantage over the original robust CS method on the original robust CS formulation, and effectively solve more sophisticated extensions where the original methods simply cannot. We demonstrate the usefulness of the extensions on several imaging tasks. |
BibTeX:
@article{pham2013efficient, author = {Pham, Duc-Son and Venkatesh, Svetha}, title = {Efficient algorithms for robust recovery of images from compressed data}, journal = {IEEE Transactions on Image Processing}, publisher = {IEEE}, year = {2013}, volume = {22}, number = {12}, pages = {4724--4737}, url = {http://arxiv.org/pdf/1211.7276.pdf} } |
Phung D, Gupta S, Nguyen T and Venkatesh S (2013), "Connectivity, online social capital, and mood: a Bayesian nonparametric analysis", IEEE Transactions on Multimedia. Vol. 15(6), pp. 1316-1325. IEEE. |
Abstract: Social capital indicative of community interaction and support is intrinsically linked to mental health. Increasing online presence is now the norm. Whilst social capital and its impact on social networks has been examined, its underlying connection to emotional response such as mood, has not been investigated. This paper studies this phenomena, revisiting the concept of “online social capital†in social media communities using measurable aspects of social participation and social support. We establish the link between online capital derived from social media and mood, demonstrating results for different cohorts of social capital and social connectivity. We use novel Bayesian nonparametric factor analysis to extract the shared and individual factors in mood transition across groups of users of different levels of connectivity, quantifying patterns and degree of mood transitions. Using more than 1.6 million users from Live Journal, we show quantitatively that groups with lower social capital have fewer positive moods and more negative moods, than groups with higher social capital. We show similar effects in mood transitions. We establish a framework of how social media can be used as a barometer for mood. The significance lies in the importance of online social capital to mental well-being in overall. In establishing the link between mood and social capital in online communities, this work may suggest the foundation of new systems to monitor online mental well-being. |
BibTeX:
@article{phung2013connectivity, author = {Phung, Dinh and Gupta, Sunil and Nguyen, Thin and Venkatesh, Svetha}, title = {Connectivity, online social capital, and mood: a Bayesian nonparametric analysis}, journal = {IEEE Transactions on Multimedia}, publisher = {IEEE}, year = {2013}, volume = {15}, number = {6}, pages = {1316--1325}, url = {http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6517534} } |
Rana S, Phung D and Venkatesh S (2013), "Split-merge augmented Gibbs sampling for hierarchical dirichlet processes", In Pacific-Asia Conference on Knowledge Discovery and Data Mining. Vol. 7819, pp. 546-557. Springer, Berlin, Heidelberg. |
Abstract: The Hierarchical Dirichlet Process (HDP) model is an important tool for topic analysis. Inference can be performed through a Gibbs sampler using the auxiliary variable method. We propose a split-merge procedure to augment this method of inference, facilitating faster convergence. Whilst the incremental Gibbs sampler changes topic assignments of each word conditioned on the previous observations and model hyper-parameters, the split-merge sampler changes the topic assignments over a group of words in a single move. This allows efficient exploration of state space. We evaluate the proposed sampler on a synthetic test set and two benchmark document corpus and show that the proposed sampler enables the MCMC chain to converge faster to the desired stationary distribution. |
BibTeX:
@inproceedings{rana2013split, author = {Rana, Santu and Phung, Dinh and Venkatesh, Svetha}, title = {Split-merge augmented Gibbs sampling for hierarchical dirichlet processes}, booktitle = {Pacific-Asia Conference on Knowledge Discovery and Data Mining}, publisher = {Springer, Berlin, Heidelberg}, year = {2013}, volume = {7819}, pages = {546--557}, url = {https://link.springer.com/chapter/10.1007/978-3-642-37456-2_46}, doi = {10.1007/978-3-642-37456-2_46} } |
Tran T, Phung D and Venkatesh S (2013), "Thurstonian Boltzmann Machines: Learning from Multiple Inequalities", In 30th International Conference on Machine Learning. , pp. 46-54. |
Abstract: We introduce Thurstonian Boltzmann Machines (TBM), a unified architecture that can naturally incorporate a wide range of data inputs at the same time. Our motivation rests in the Thurstonian view that many discrete data types can be considered as being generated from a subset of underlying latent continuous variables, and in the observation that each realisation of a discrete type imposes certain inequalities on those variables. Thus learning and inference in TBM reduce to making sense of a set of inequalities. Our proposed TBM naturally supports the following types: Gaussian, intervals, censored, binary, categorical, muticategorical, ordinal, (in)-complete rank with and without ties. We demonstrate the versatility and capacity of the proposed model on three applications of very different natures; namely handwritten digit recognition, collaborative filtering and complex social survey analysis. |
BibTeX:
@inproceedings{tran2013thurstonian, author = {Tran, Truyen and Phung, Dinh and Venkatesh, Svetha}, title = {Thurstonian Boltzmann Machines: Learning from Multiple Inequalities}, booktitle = {30th International Conference on Machine Learning}, year = {2013}, pages = {46--54}, url = {http://jmlr.csail.mit.edu/proceedings/papers/v28/tran13.pdf} } |
Tran T, Phung D, Luo W, Harvey R, Berk M and Venkatesh S (2013), "An integrated framework for suicide risk prediction", In The 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. , pp. 1410-1418. |
Abstract: Suicide is a major concern in society. Despite of great attention paid by the community with very substantive medico-legal implications, there has been no satisfying method that can reliably predict the future attempted or completed suicide. We present an integrated machine learning framework to tackle this challenge. Our proposed framework consists of a novel feature extraction scheme, an embedded feature selection process, a set of risk classifiers and finally, a risk calibration procedure. For temporal feature extraction, we cast the patient's clinical history into a temporal image to which a bank of one-side filters are applied. The responses are then partly transformed into mid-level features and then selected in l1-norm framework under the extreme value theory. A set of probabilistic ordinal risk classifiers are then applied to compute the risk probabilities and further re-rank the features. Finally, the predicted risks are calibrated. Together with our Australian partner, we perform comprehensive study on data collected for the mental health cohort, and the experiments validate that our proposed framework outperforms risk assessment instruments by medical practitioners. |
BibTeX:
@inproceedings{TranPLHBV13, author = {Truyen Tran and Dinh Phung and Wei Luo and Richard Harvey and Michael Berk and Svetha Venkatesh}, title = {An integrated framework for suicide risk prediction}, booktitle = {The 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining}, year = {2013}, pages = {1410--1418}, url = {http://prada-research.net/ truyen/papers/KDD-published.pdf} } |
Venkatesh S, Phung D, Tran T and Gupta S (2013), "Capitalising on the data deluge: Data analytics for healthcare".
[BibTeX] |
BibTeX:
@misc{Venkatesh_etal_13capitalising, author = {S. Venkatesh and D. Phung and T. Tran and S. Gupta}, title = {Capitalising on the data deluge: Data analytics for healthcare}, booktitle = {Big Data. Health Informatics Society of Australia (HISA)}, year = {2013} } |
Venkatesh S, Greenhill S, Phung D, Duong T, Adams B, Marshall W and Cairns D (2013), "TOBY: Therapy Outcomes By You", In Association for Behavior Analysis International (ABAI)., Association for Behavior Analysis International. Portland, Oregon, USA , pp. poster session 2. |
Abstract: Early intervention is critical for children diagnosed with autism. Unfortunately, there is often a long gap of waiting, and wasting, time between a "formal" diagnosis and therapy. We describe TOBY Playpad (www.tobyplaypad.com) whose goal is to close this gap by empowering parents to help their children early. TOBY stands for Therapy Outcome by You and currently is an iPad application. It provides an adaptive syllabus of more than 320 activities developed by autism and machine learning experts to target key development areas which are known to be deficit for ASD children such as imitation, joint attention and language. TOBY delivers lessons, materials, instructions and interactions for both on-iPad and Natural Environment Tasks (NET) off-iPad activities. TOBY is highly adaptive and personalized, intelligently increasing its complexity, varying prompts and reinforcements as the child progresses over time. Prompting and reinforcing strategies are also recommended for parents to make the most of everyday opportunities to teach children. Essentially, TOBY removes the burden on parents from extensive preparation of materials and manual data recording. Three trials on 20, 50 and 36 children with AutismWest (www.autismwest.org.au) have been conducted since last year. The results are promising providing evidence of learning shown in skills that were not present previously in some children. NET activities are shown to be effective for children and popular with parents. |
BibTeX:
@inproceedings{Venkatesh_etal_ABAI13, author = {S. Venkatesh and S. Greenhill and D. Phung and T. Duong and B. Adams and W. Marshall and D. Cairns}, title = {TOBY: Therapy Outcomes By You}, booktitle = {Association for Behavior Analysis International (ABAI)}, year = {2013}, pages = {poster session 2}, url = {2013/conferences/Venkatesh_etal_ABAI13.pdf} } |
Venkatesh S, Phung D, Duong T, Greenhill S and Adams B (2013), "TOBY: early intervention in autism through technology", In SIGCHI Conference on Human Factors in Computing Systems. , pp. 3187-3196. |
Abstract: We describe TOBY Playpad, an early intervention program for children with Autism Spectrum Disorder (ASD). TOBY teaches the teacher -- the parent -- during the crucial period following diagnosis, which often coincides with no access to formal therapy. We reflect on TOBY's evolution from table-top aid for flashcards to an iPad app covering a syllabus of 326 activities across 51 skills known to be deficient for ASD children, such imitation, joint attention and language. The design challenges unique to TOBY are the need to adapt to marked differences in each child's skills and rate of development (a trait of ASD) and teach parents unfamiliar concepts core to behavioural therapy, such as reinforcement, prompting, and fading. We report on three trials that successively decrease oversight and increase parental autonomy, and demonstrate clear evidence of learning. TOBY's uniquely intertwined Natural Environment Tasks are found to be effective for children and popular with parents. |
BibTeX:
@inproceedings{Venkatesh2013, author = {Venkatesh, Svetha and Phung, Dinh and Duong, Thi and Greenhill, Stewart and Adams, Brett}, title = {TOBY: early intervention in autism through technology}, booktitle = {SIGCHI Conference on Human Factors in Computing Systems}, year = {2013}, pages = {3187--3196}, url = {http://dl.acm.org/citation.cfm?id=2466437} } |