Breadcrumb
3D motion tracking of the heart using Harmonic Phase (HARP) isosurfaces
Tags are non-invasive features induced in the heart muscle that enable the tracking of heart motion. Each tag line, in fact, corresponds to a 3D tag surface that deforms with the heart muscle during the cardiac cycle. Tracking of tag surfaces deformation is useful for the analysis of left ventricular motion. Cardiac material markers (Kerwin et al, MIA, 1997) can be obtained from the intersections of orthogonal surfaces which can be reconstructed from short- and long-axis tagged images. The proposed method uses Harmonic Phase (HARP) method for tracking tag lines corresponding to a specific
TopicAnalyzer: A system for unsupervised multi-label Arabic topic categorization
The wide spread use of social media tools and forums has led to the production of textual data at unprecedented rates. Without summarization, classification or other form of analysis, the sheer volume of this data will often render it useless and human analysis on this scale is next to impossible. The work presented in this paper focuses on investigating an approach for classifying large volumes of data when no training data and no classification scheme are available. Motivation for this work lies in encountering a real life problem which is further described in the paper. The presented system
Person name extraction from modern standard Arabic or colloquial text
Person Name extraction from Arabic text is a challenging task. While most existing Arabic texts are written in Modern Standard Arabic Text (MSA) the volume of Arabic Colloquial text is increasing progressively with the wide spread use of social media examples of which are Facebook, Google Moderator and Twitter. Previous work addressed extracting persons' names from MSA text only and especially from news articles. Previous work also relied on a lot of resources such as gazetteers for places, organizations, verbs, and person names. In this paper we introduce a system for extracting persons'
A novel algorithm for simultaneous face detection and recognition
Face detection and recognition has been introduced in many real world applications. Several algorithms have been implemented for either detection or recognition. In this paper, a novel algorithm, which simultaneously detects and recognizes facial images employing the same method, is presented. The proposed algorithm is based on a new 2D representation for the Histogram of Oriented Gradients (2D-HOG) in conjunction with 2DPCA for feature extraction. Experimental results conducted on existing datasets, FERET, ORL, UMIST, JAFFE, and MIT-CMU, achieved better accuracy and running time compared with
Quantitative assessment of Diabetic Macular Edema using fluorescein leakage maps
Diagnosis of Diabetic Macular Edema (DME) from Fundus Fluorescein Angiography (FFA) image sequences is a standard clinical practice. Nevertheless, current methods depend on subjective evaluation of the amount of fluorescein leakage in the images which lack reproducibility and require well-trained grader. In this work, we present a method for processing FFA images to generate a fluorescein leakage map that can be used for quantitative evaluation of DME. An essential step in the proposed method is to model the spatial distribution of the image intensity within the macula. This model, which
Quantitative assessment of age-related macular degeneration using parametric modeling of the leakage transfer function: Preliminary results
Age-related macular degeneration (AMD) is a major cause of blindness and visual impairment in older adults. The wet form of the disease is characterized by abnormal blood vessels forming a choroidal neovascular membrane (CNV), that result in destruction of normal architecture of the retina. Current evaluation and follow up of wet AMD include subjective evaluation of Fluorescein Angiograms (FA) to determine the activity of the lesion and monitor the progression or regression of the disease. However, this subjective evaluation prevents accurate monitoring of the disease progression or regression
An approach for deriving semantically related category hierarchies from Wikipedia category graphs
Wikipedia is the largest online encyclopedia known to date. Its rich content and semi-structured nature has made it into a very valuable research tool used for classification, information extraction, and semantic annotation, among others. Many applications can benefit from the presence of a topic hierarchy in Wikipedia. However, what Wikipedia currently offers is a category graph built through hierarchical category links the semantics of which are un-defined. Because of this lack of semantics, a sub-category in Wikipedia does not necessarily comply with the concept of a sub-category in a
Plagiarism candidate retrieval using selective query formulation and discriminative query scoring: Notebook for PAN at CLEF 2013
This paper details the approach of implementing an English plagiarism source retrieval system to be presented at PAN 2013. The system uses the TextTiling algorithm to break a given document into segments that are centered around certain topics within the document. From these segments, keyphrases are generated using the KPMiner keyphrase extraction system. These keyphrases and segments are then used in generating queries indicative of the segment, and consequently the document. The queries are submitted to ChatNoir for finding plagiarism sources in the ClueWeb09 corpus from which the pan13
2DHOOF-2DPCA contour based optical flow algorithm for human activity recognition
A novel algorithm for human activity recognition is presented in this paper. This approach is based on a new 2D representation for the Histogram of Oriented Optical Flow (2DHOOF) describing the motion of the actor's contour, where one multi-layer 2D-histogram per video is constructed. Each histogram layer consists of 2D bins (layers) that represent different range of angles. Applying our 2DHOOF features descriptors on the actor's contour reduces the storage requirement and the computation complexity since a sparse optical flow is calculated instead of dense optical flow. In addition, it is
A collaborative resource to build consensus for automated left ventricular segmentation of cardiac MR images
A collaborative framework was initiated to establish a community resource of ground truth segmentations from cardiac MRI. Multi-site, multi-vendor cardiac MRI datasets comprising 95 patients (73 men, 22 women; mean age 62.73. ±. 11.24. years) with coronary artery disease and prior myocardial infarction, were randomly selected from data made available by the Cardiac Atlas Project ( Fonseca et al., 2011). Three semi- and two fully-automated raters segmented the left ventricular myocardium from short-axis cardiac MR images as part of a challenge introduced at the STACOM 2011 MICCAI workshop (
Pagination
- Previous page ‹‹
- Page 8
- Next page ››